orionweller commited on
Commit
93d5d7d
1 Parent(s): 3fb9e75

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +30 -0
  2. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  3. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  4. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds +3 -0
  5. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds +3 -0
  6. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds +3 -0
  7. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds +3 -0
  8. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  9. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  10. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  11. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  12. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  13. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  14. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  15. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds +3 -0
  16. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  17. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  18. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds +3 -0
  19. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  20. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds +3 -0
  21. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds +3 -0
  22. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds +3 -0
  23. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds +3 -0
  24. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds +3 -0
  25. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds +3 -0
  26. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds +3 -0
  27. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds +3 -0
  28. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds +3 -0
  29. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds +3 -0
  30. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  31. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds +3 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10401-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10401-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12012-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12012-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13602-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13602-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17071-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17071-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17075-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17075-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20365-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20365-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21415-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21415-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22067-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22067-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25489-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25489-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27161-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
.gitattributes CHANGED
@@ -11112,3 +11112,33 @@ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/
11112
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11113
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11114
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11112
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11113
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11114
  train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
11115
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11116
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds filter=lfs diff=lfs merge=lfs -text
11117
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
11118
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds filter=lfs diff=lfs merge=lfs -text
11119
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
11120
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text
11121
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds filter=lfs diff=lfs merge=lfs -text
11122
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
11123
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text
11124
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11125
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
11126
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
11127
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11128
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
11129
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds filter=lfs diff=lfs merge=lfs -text
11130
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text
11131
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11132
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text
11133
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
11134
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
11135
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text
11136
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
11137
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text
11138
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11139
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds filter=lfs diff=lfs merge=lfs -text
11140
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11141
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
11142
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11143
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11144
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:197c46edf98aa6c44fd5658c7b98b789595b1d85c5352efc2c14afe29236cbce
3
+ size 67108782
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace2f29f3d1d81d733c216f80e6e5cd3f42014bfaee9358151a1306141ba60c3
3
+ size 67107934
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe3f3fed1e10aea36b505795669b77bf317b7fb21e4b1cc80a3149982fae4c46
3
+ size 67107846
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca831ab18814eadb632cdd5fa7746499ac2904ea2024b4336769b9f1abe0985a
3
+ size 67107700
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dcb2dcc87941dda8fce1be852ce484f9fb4adacc9e5612c0511ba1ce9ebfc80
3
+ size 67108156
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c10ddf70289acfa0f2a63d002e0bbfc11e73f172c32b2cf2babe1fb28bfae0d
3
+ size 67107979
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ec2ef3db3999f08efd9d332f49ce926eb4da71d8363a166c9893abe6b8e7866
3
+ size 67107632
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:162c243029298d345419a921a470945d6a4c67c1b62356da4f77fbb3580c9d0b
3
+ size 67108496
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb49a48e82e3233b9cbbbad6128956dad03b1b707016d1479827ecebe567e153
3
+ size 67108594
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eb616aa970c050f681f1759e474f3f860196528d70a87016027417f37927603
3
+ size 67107778
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49cf23859ac59b38ab0b71218c59b8720b2f09a2b42d8cec6d999b48b4b349b1
3
+ size 67108413
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffe0b48a6992d94fafb4b68c44f49ad6e0016694174121225397387195406809
3
+ size 67107859
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161a7ceac9922945bc29dc2cbc608c41feb8859cc034a47ef7cc662b00ed22d8
3
+ size 67107940
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8261b687e2d18abf0da138b68f3a162cf7cd62534aba889633ac22ca9ad9820
3
+ size 67106898
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54d1e9f89d671a743f126d8b7a7ab294ebac5eccc38bd2e967d9dc8b6a5fefbe
3
+ size 67106975
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:129cd2ccb0d98406f9b58a3aadd901c76a6006417dbf801e4ece412369d0304b
3
+ size 67108049
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69ff381b71a9b4115ba8dc690a1010fd899ea3365e47c5840a76264f2b562f9d
3
+ size 67108122
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ea9b14326efc2f72721a0db3f26968cda8ae368b82385d82f4a49b5e6f042ed
3
+ size 67108844
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3d597518f11a2e9b5d4e2aec4d31a33e6ae6bc22f941d74013040bcb9c7a943
3
+ size 67107665
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bccb687ca6224dbf16c1bc8b258bb60c9e10d1fe3682ce593dc71d2b3c74cb9
3
+ size 67107831
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6e5879f78c50de2e76c7e12c1f195bb5a754d1a97513d64a3598c12cd7c9733
3
+ size 67107279
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6256cc52754c4ceb9491caa07488434f3a81ce37add5696ff22727742184e1d
3
+ size 67107622
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5516b6538d073c1ba5416c28d9e22215e7526dcd361328619680140058136848
3
+ size 67108723
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef785fd78718bfd1c5f4d0ff4c52e2654e5b08045ef741c9c0f285331aa3b95f
3
+ size 67108530
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62f4c21906fcf56397bfedad8aa1ead72df361fe65c10ac0328e83af955bbe87
3
+ size 67106789
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56247aa7f526c4a2d487454b28169c1a1d5c5556b0dd733a051eb0f71890523c
3
+ size 67107944
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fceda0b48512f9efccaa4d3b34cd85772ee14bc753ce90a81d3d605185d79fa6
3
+ size 67107467
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8cf4759ff3990c347237045b5c462ea71bb4087cf7a73493a9f5babd8c36f28
3
+ size 67107952
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14440466292296292658b54d7305b14b27f3318a65e3d320b0df40b6754d51b3
3
+ size 67108438
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f933e5f0aeb5efb31f48ed7c67f9f33012248923ee00cc9f57e883789772c773
3
+ size 56741207
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10401-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107636, "hashes": {}}, "samples": 42206, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47578698, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22552065, "hashes": {}}, "samples": 14626, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16102588, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10401-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 43070594,
3
+ "num_truncated_tokens": 43033029
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12012-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107654, "hashes": {}}, "samples": 43819, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47820739, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14679577, "hashes": {}}, "samples": 9417, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10428080, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12012-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39245943,
3
+ "num_truncated_tokens": 39214142
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13602-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108228, "hashes": {}}, "samples": 43648, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47504834, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13700014, "hashes": {}}, "samples": 8858, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9727703, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13602-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38778961,
3
+ "num_truncated_tokens": 38748812
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17071-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107450, "hashes": {}}, "samples": 43820, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47429900, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15843554, "hashes": {}}, "samples": 10209, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11223108, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17071-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39803484,
3
+ "num_truncated_tokens": 39770118
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17075-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108836, "hashes": {}}, "samples": 43464, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47511819, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17288760, "hashes": {}}, "samples": 11049, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12247817, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17075-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40511630,
3
+ "num_truncated_tokens": 40477723
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20365-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108778, "hashes": {}}, "samples": 43748, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47766893, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14714482, "hashes": {}}, "samples": 9349, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10421153, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20365-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39268242,
3
+ "num_truncated_tokens": 39236470
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21415-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107464, "hashes": {}}, "samples": 43214, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47605429, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18392630, "hashes": {}}, "samples": 11764, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13002015, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21415-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41048337,
3
+ "num_truncated_tokens": 41013658
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22067-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106946, "hashes": {}}, "samples": 43263, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47981410, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15686206, "hashes": {}}, "samples": 10264, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11224932, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22067-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39739702,
3
+ "num_truncated_tokens": 39706673
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25489-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108208, "hashes": {}}, "samples": 43984, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48012898, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12841654, "hashes": {}}, "samples": 8342, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9194169, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25489-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38355496,
3
+ "num_truncated_tokens": 38324115
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27161-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108138, "hashes": {}}, "samples": 43591, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47983383, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13214760, "hashes": {}}, "samples": 8764, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9492021, "hashes": {}}}], "version": 2}