orionweller commited on
Commit
c4d9660
1 Parent(s): a843fd3

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +34 -0
  2. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  3. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  4. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  5. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  6. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  7. train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  8. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  9. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  10. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  11. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  12. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  13. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  14. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  15. train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  16. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  17. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  18. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  19. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  20. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  21. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  22. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  23. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  24. train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  25. train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  26. train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  27. train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  28. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  29. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  30. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  31. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  32. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  33. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  34. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  35. train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11347-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11347-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11586-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11586-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13588-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13588-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14213-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14213-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15618-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15618-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1599-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1599-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17178-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17178-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18676-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
.gitattributes CHANGED
@@ -12877,3 +12877,37 @@ train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
12877
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12878
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12879
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12877
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12878
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12879
  train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
12880
+ train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12881
+ train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12882
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12883
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12884
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12885
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12886
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12887
+ train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12888
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12889
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12890
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12891
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12892
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12893
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12894
+ train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12895
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12896
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12897
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12898
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12899
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12900
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12901
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12902
+ train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12903
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12904
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12905
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12906
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12907
+ train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12908
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12909
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12910
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12911
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12912
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12913
+ train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f878d62741af38d7d1f2f0cf456fc21bcd999ad874999a6639f94afa2f60dd1
3
+ size 67107455
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df34a7a12497a8573fbe42515005809e92666633faac648a7d2a409dfa607fdd
3
+ size 67107594
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f94227fa21e48e00f0fff27ad8806df0feb200c2a2c25f44120c2207720d205a
3
+ size 67107872
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f4b68f96d44c1f9e6e48629512103cad16462edfcf79a280c84fcc97a60c782
3
+ size 67107139
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70acedbd15137b7805904c92007fdcdf34ce67fe44c6037392ad609da97f13cb
3
+ size 67108123
train/arxiv/arxiv_0009-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d6cf7de6418692f438096dd38536138a3e14bba3c1765a5ea5adf86e9bfe088
3
+ size 15833528
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:734ed89e9d6f5d604289a44f0933fa97a0ba4268d55d15939196d493fd1d0e1e
3
+ size 67107989
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03a835f89952bffee7d2d7f258b5e0d81937ac6f8849f7f9c8d6b955c53c048d
3
+ size 67107851
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dda623c397ea5ee62096f1ffa5c0da4d129f0239c8f1bb8aed842b2e60101779
3
+ size 67107580
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1be4b81073852e9f4ed29c7f05aa2e3a9430aafd591b9547ddef41004b40c7cd
3
+ size 67107336
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed94add882ebbbea8916f5e1581c38a3183ce5f5a3002157c259396ce5ec5be
3
+ size 67108050
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f9e7d3767b07bcd168e558579643b684ced861dd083c4134806b9e106c84470
3
+ size 67107539
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5efa3d80cf8deeee244bfba9305beef93a7f1ee59f21dfc27c72f5eb9c846601
3
+ size 67107195
train/arxiv/arxiv_0027-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e1179c5c5c03eaae7df461341c9a151c0ad8756e50a2be87d56b157680893b4
3
+ size 48193061
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28e67b712ed446379669738b011cec64f0977d5e069850cd2a72871b85ab6050
3
+ size 67107866
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8160f649d48d2033e59235b8bcbfce98f211f262604b6d56ef4ce27467ce2fb
3
+ size 67108564
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46ece38ae42c590b92e1de0665261e0270de620f7ec3b85baa642bf2f1d02e13
3
+ size 67108788
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72d595295065e01329001338f2e14e04769a683a02bcdf93a559980c5e07c72d
3
+ size 67107552
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c95c74147089e1612db01789977efe3423fa53e10e4b258d0770403667d72347
3
+ size 67107446
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46b2346ca9c9be072c9e4dc2f620b96cbd6d5422c362c6233254a947db14f931
3
+ size 67107934
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7acf3372c76870b162702ece09ad0bcd7bfa8fb0dae3ec2dc2a98646e914d94c
3
+ size 67107392
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d486dd70d328d689c5bcd880de4fe99c566445c5d3d777ce2259de29e7e4237
3
+ size 67106814
train/arxiv/arxiv_0048-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4aaa6ba7718d998193bdd92d2a29066e8f7e5a3d4c5be31e3d5afce05c0c625
3
+ size 21496964
train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec5b8c43bab41726a968b331fde0d9f17de20a39f4f5f3e30ce91d4e2f16080b
3
+ size 67107265
train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ba59007a35d40b5698b63e0eb04a114d8e2348a5e177c67a09333401f288132
3
+ size 67108015
train/arxiv/arxiv_0053-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9a0df3993dd4f6ddd0d94a911c24247fbbb0770a57e016c9410128458be350
3
+ size 67108277
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0786fc9997427e23ed31434d7c62f85f3b54643828d29afa39f91f72f08daa8
3
+ size 67107813
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98a867795d9090383f4d5a53cec36c532d92a67973003423e5e86eb353d131ba
3
+ size 67108529
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae4b13a2bfb703909f21c849738206ef2ef97ea3ea02000e4ea3e187b1631a1f
3
+ size 67107218
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e90e568165bf8e61a251afc47da4f7b7a99e136e3e438d0f9f29b5232dbacfa1
3
+ size 67108618
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c95445c13b66d68299ba47a36185666a30d43689922199cd83c29e8965751276
3
+ size 67108164
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bc7275b5e8c5cd7ba866cb02b5b5459e16cc95be921910a7a39c92cc632601f
3
+ size 67106773
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f390727925826a901e0310c4f3007a8071d27f6d69af9287da0720902771509
3
+ size 67107670
train/arxiv/arxiv_0071-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803e342d14ca6cd92d1dc4b15b28dc9d04096998c9a6e7b92af6e5026e2e6f8d
3
+ size 47769405
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11347-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108860, "hashes": {}}, "samples": 43515, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47517053, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17248505, "hashes": {}}, "samples": 10914, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12201917, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11347-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40493971,
3
+ "num_truncated_tokens": 40459906
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11586-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107367, "hashes": {}}, "samples": 43453, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47700649, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13627033, "hashes": {}}, "samples": 9176, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9738219, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11586-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38738319,
3
+ "num_truncated_tokens": 38707585
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13588-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108415, "hashes": {}}, "samples": 43982, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47402596, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11007107, "hashes": {}}, "samples": 7153, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7800803, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13588-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37475036,
3
+ "num_truncated_tokens": 37447092
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14213-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106790, "hashes": {}}, "samples": 44032, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47730816, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10699675, "hashes": {}}, "samples": 7099, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7614427, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14213-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37320746,
3
+ "num_truncated_tokens": 37292163
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15618-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108829, "hashes": {}}, "samples": 42306, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47703011, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22159119, "hashes": {}}, "samples": 14344, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15815988, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15618-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42880335,
3
+ "num_truncated_tokens": 42842254
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1599-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106946, "hashes": {}}, "samples": 42579, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47602947, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21423561, "hashes": {}}, "samples": 13660, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15263419, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1599-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42524338,
3
+ "num_truncated_tokens": 42487461
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17178-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108232, "hashes": {}}, "samples": 42887, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47473945, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19835488, "hashes": {}}, "samples": 12649, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14030401, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17178-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41752774,
3
+ "num_truncated_tokens": 41717100
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18676-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107702, "hashes": {}}, "samples": 44104, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47883414, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12188233, "hashes": {}}, "samples": 8148, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8657153, "hashes": {}}}], "version": 2}