orionweller commited on
Commit
7d912be
·
verified ·
1 Parent(s): 23c952e

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. train/dclm-sampled/index.json +0 -0
  3. train/dclm-sampled/index.json.old +3 -0
  4. train/dclm-sampled/index.json.old_before_downsample +3 -0
  5. train/dclm-sampled/num_tokens.json +4 -0
  6. train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  7. train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  8. train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  9. train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  10. train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  11. train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  12. train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  13. train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  14. train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  15. train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  16. train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  17. train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  18. train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  19. train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  20. train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  21. train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  22. train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  23. train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  24. train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  25. train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  26. train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  27. train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  28. train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  29. train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  30. train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  31. train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  32. train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  33. train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  34. train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  35. train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  36. train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  38. train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  39. train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  40. train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  41. train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  43. train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  44. train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  45. train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
  46. train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  48. train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  49. train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
  50. train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
.gitattributes CHANGED
@@ -117,3 +117,5 @@ train/stackexchange/stackexchange_0001-tokenized-chunked-1024-512-128-backfill-n
117
  train/books/books_0000-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json filter=lfs diff=lfs merge=lfs -text
118
  train/books/books_0001-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json filter=lfs diff=lfs merge=lfs -text
119
  train/fineweb2/index.json filter=lfs diff=lfs merge=lfs -text
 
 
 
117
  train/books/books_0000-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json filter=lfs diff=lfs merge=lfs -text
118
  train/books/books_0001-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json filter=lfs diff=lfs merge=lfs -text
119
  train/fineweb2/index.json filter=lfs diff=lfs merge=lfs -text
120
+ train/dclm-sampled/index.json.old filter=lfs diff=lfs merge=lfs -text
121
+ train/dclm-sampled/index.json.old_before_downsample filter=lfs diff=lfs merge=lfs -text
train/dclm-sampled/index.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/index.json.old ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6e1fc73f4e0d7f434f251a881089d63f58ae953046f795ca4dd3def4368d6a
3
+ size 15813921
train/dclm-sampled/index.json.old_before_downsample ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6e1fc73f4e0d7f434f251a881089d63f58ae953046f795ca4dd3def4368d6a
3
+ size 15813921
train/dclm-sampled/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 200257531665,
3
+ "num_skipped_tokens": 634591723
4
+ }
train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107313, "hashes": {}}, "samples": 44776, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48833369, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 7346646, "hashes": {}}, "samples": 4914, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5343781, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4039e30fdb92e14e9c5ecd5334ae546e552ae2d427138ebae06b20b16a59f6b4
3
+ size 67107313
train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ebb769f36781ca5a4b6b83cfc39c130c162e3853c2a3123d9177b98dafab02
3
+ size 7346646
train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 35689209, "total_tokens_skipped": 125661, "percentiles": {"0th": 26, "10th": 230, "20th": 395, "30th": 538, "40th": 657, "50th": 793, "60th": 939, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_10147-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107335, "hashes": {}}, "samples": 42850, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48191555, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18888502, "hashes": {}}, "samples": 12399, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13671262, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14beeebc9766b282fe4f0db7fb3846ccdf8f968139c9e78ed1d45280d581c8f4
3
+ size 67107335
train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff81a923d5e674d2708019744161e3319353d6fc31ea5bd239f38a081d058eeb
3
+ size 18888502
train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 41287852, "total_tokens_skipped": 126970, "percentiles": {"0th": 28, "10th": 249, "20th": 421, "30th": 572, "40th": 714, "50th": 871, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_14334-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108020, "hashes": {}}, "samples": 44740, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48608763, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9723673, "hashes": {}}, "samples": 6458, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7037390, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6347a8c4c32e389e32986ed5abd4c347cb3316988ebc456e416a05e76b7e2901
3
+ size 67108020
train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96cead5a8a7130e6c63b5c7e9b751b5314898bfd23b3201bbe716a892c4ba183
3
+ size 9723673
train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36831564, "total_tokens_skipped": 117132, "percentiles": {"0th": 19, "10th": 220, "20th": 387, "30th": 534, "40th": 658, "50th": 801, "60th": 963, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_19126-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108161, "hashes": {}}, "samples": 43316, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48034247, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18009872, "hashes": {}}, "samples": 11735, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12953750, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b06e0df157efbfc95a636b5eb8fbca17f9612611b2f41c81d8da713a6387b8d6
3
+ size 67108161
train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75c220bfff34897664ca13c56db3a6d781aea47e6e458b3102ee32f1811b569c
3
+ size 18009872
train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 40855309, "total_tokens_skipped": 126608, "percentiles": {"0th": 27, "10th": 234, "20th": 410, "30th": 564, "40th": 708, "50th": 864, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_20611-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108686, "hashes": {}}, "samples": 44651, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48423616, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10501605, "hashes": {}}, "samples": 6681, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7548842, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42f66eba8e6f72860d326c1a0f5929054551d6fc9c0e88a98c9e6e8460c0180a
3
+ size 67108686
train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc4e72ab3bfbf108f439ebc70b0da18c3905169a987ffcd166545ef3775ad6f
3
+ size 10501605
train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37216574, "total_tokens_skipped": 124908, "percentiles": {"0th": 24, "10th": 232, "20th": 395, "30th": 543, "40th": 672, "50th": 812, "60th": 969, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_22066-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107518, "hashes": {}}, "samples": 42743, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48294969, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21218305, "hashes": {}}, "samples": 13745, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15236967, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e35e1d4d1d6ab8ad0442bf1f82d9b84f40d5f5ce3dfe051b51379350961d67e0
3
+ size 67107518
train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:285a45bf3b4d5b0d27a2000f397e56bfce4634626a8dcbca1fd584cbdef311d6
3
+ size 21218305
train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 42414477, "total_tokens_skipped": 124648, "percentiles": {"0th": 25, "10th": 250, "20th": 423, "30th": 575, "40th": 722, "50th": 886, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_23768-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107624, "hashes": {}}, "samples": 42992, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47763750, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19008409, "hashes": {}}, "samples": 12319, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13672780, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15962d5a34e5ca75e968d58fe34717b89674103798d3545927637b43dc9953c4
3
+ size 67107624
train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:591241041b6cc36e4038acaec9dcea53e1cf1cbf4e1957baf0504daa834f5601
3
+ size 19008409
train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 41346049, "total_tokens_skipped": 134639, "percentiles": {"0th": 32, "10th": 247, "20th": 423, "30th": 575, "40th": 713, "50th": 869, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_2543-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108681, "hashes": {}}, "samples": 44075, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48833169, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12600578, "hashes": {}}, "samples": 8355, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9149266, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b67b5c41fb8f62f81efd7f3fc9107502387081c32fd46ecc15d9382983c915cf
3
+ size 67108681
train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bbf33d3072deee16a2f16150f050b94b7f20d8fe107ba2ce745487f1355fe43
3
+ size 12600578
train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38231996, "total_tokens_skipped": 116408, "percentiles": {"0th": 26, "10th": 236, "20th": 400, "30th": 545, "40th": 673, "50th": 821, "60th": 991, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_37784-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107445, "hashes": {}}, "samples": 44845, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48769433, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8039948, "hashes": {}}, "samples": 5292, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5841166, "hashes": {}}}], "version": 2}
train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:238f977222c2138132e1978a66cf99f269a8460d2a0ac7d2334040a2df4ba080
3
+ size 67107445
train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36935a942776d544a575d7883d5ca2e39b8c4e36685299367c2582279be7750e
3
+ size 8039948
train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36022183, "total_tokens_skipped": 122295, "percentiles": {"0th": 31, "10th": 225, "20th": 392, "30th": 537, "40th": 658, "50th": 794, "60th": 947, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
train/dclm-sampled/split_42197-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff