mtasic85 commited on
Commit
286a68d
·
1 Parent(s): fdc8b84

prepare datasets core

Browse files
Files changed (1) hide show
  1. README.md +24 -43
README.md CHANGED
@@ -53,42 +53,38 @@ time python -B prepare_core_datasets.py
53
  ```
54
 
55
  ```
56
- i=0, min_len=0, max_len=1073741824, block_size=1025, chunk_size=16400000, len(dataset)=5146620, len(dataset) * block_size=5275285500
57
- Total number of tokens in the optimized dataset '../core-data-0-0-1073741824-1025-16000' is 5275285500
58
 
59
- i=1, min_len=1025, max_len=2049, block_size=2049, chunk_size=16392000, len(dataset)=309838, len(dataset) * block_size=634858062
60
- Total number of tokens in the optimized dataset '../core-data-1-1025-2049-2049-8000' is 634858062
61
 
62
- i=2, min_len=2049, max_len=4097, block_size=4097, chunk_size=16388000, len(dataset)=113843, len(dataset) * block_size=466414771
63
- Total number of tokens in the optimized dataset '../core-data-2-2049-4097-4097-4000' is 466414771
64
 
65
- i=3, min_len=4097, max_len=8193, block_size=8193, chunk_size=16386000, len(dataset)=56713, len(dataset) * block_size=464649609
66
- Total number of tokens in the optimized dataset '../core-data-3-4097-8193-8193-2000' is 464649609
67
 
68
- i=4, min_len=8193, max_len=16385, block_size=16385, chunk_size=16385000, len(dataset)=37406, len(dataset) * block_size=612897310
69
- Total number of tokens in the optimized dataset '../core-data-4-8193-16385-16385-1000' is 612897310
70
 
71
- i=5, min_len=16385, max_len=32769, block_size=32769, chunk_size=16384500, len(dataset)=12737, len(dataset) * block_size=417378753
72
- Total number of tokens in the optimized dataset '../core-data-5-16385-32769-32769-500' is 417378753
73
 
74
- i=6, min_len=32769, max_len=65537, block_size=65537, chunk_size=16384250, len(dataset)=2824, len(dataset) * block_size=185076488
75
- Total number of tokens in the optimized dataset '../core-data-6-32769-65537-65537-250' is 185076488
76
 
77
- i=7, min_len=65537, max_len=131073, block_size=131073, chunk_size=16384125, len(dataset)=634, len(dataset) * block_size=83100282
78
- Total number of tokens in the optimized dataset '../core-data-7-65537-131073-131073-125' is 83100282
79
 
80
- real 292m54.341s
81
- user 2118m1.154s
82
- sys 12m2.746s
83
-
84
- 20G tangled-alpha-0.9-core/core-data-0-0-1073741824-1025-16000
85
- 2.4G tangled-alpha-0.9-core/core-data-1-1025-2049-2049-8000
86
- 1.8G tangled-alpha-0.9-core/core-data-2-2049-4097-4097-4000
87
- 1.8G tangled-alpha-0.9-core/core-data-3-4097-8193-8193-2000
88
- 2.3G tangled-alpha-0.9-core/core-data-4-8193-16385-16385-1000
89
- 1.6G tangled-alpha-0.9-core/core-data-5-16385-32769-32769-500
90
- 709M tangled-alpha-0.9-core/core-data-6-32769-65537-65537-250
91
- 321M tangled-alpha-0.9-core/core-data-7-65537-131073-131073-125
92
  ```
93
 
94
  ```bash
@@ -96,21 +92,6 @@ CUDA_VISIBLE_DEVICES=0 CUDA_LAUNCH_BLOCKING=0 PYTORCH_CUDA_ALLOC_CONF=expandable
96
  ```
97
 
98
  ```
99
- Seed set to 23
100
- Time to instantiate model: 0.21 seconds.
101
- Total parameters: 302,023,168
102
- Verifying settings ...
103
- Measured TFLOPs: 55520.94
104
- Epoch 1 | iter 64 step 1 | loss train: 11.982, val: n/a | iter time: 409.55 ms (step) remaining time: 4 days, 17:45:21
105
- Epoch 1 | iter 128 step 2 | loss train: 11.980, val: n/a | iter time: 354.46 ms (step) remaining time: 3 days, 15:01:16
106
- Epoch 1 | iter 192 step 3 | loss train: 11.980, val: n/a | iter time: 353.67 ms (step) remaining time: 3 days, 5:46:03
107
- Epoch 1 | iter 256 step 4 | loss train: 11.980, val: n/a | iter time: 354.11 ms (step) remaining time: 3 days, 1:05:26
108
- Epoch 1 | iter 320 step 5 | loss train: 11.978, val: n/a | iter time: 358.28 ms (step) remaining time: 2 days, 22:21:45
109
- Epoch 1 | iter 384 step 6 | loss train: 11.974, val: n/a | iter time: 356.21 ms (step) remaining time: 2 days, 20:33:55
110
- Epoch 1 | iter 448 step 7 | loss train: 11.964, val: n/a | iter time: 357.42 ms (step) remaining time: 2 days, 19:15:59
111
- Epoch 1 | iter 512 step 8 | loss train: 11.956, val: n/a | iter time: 355.74 ms (step) remaining time: 2 days, 18:16:43
112
- Epoch 1 | iter 576 step 9 | loss train: 11.937, val: n/a | iter time: 356.05 ms (step) remaining time: 2 days, 17:28:34
113
- Epoch 1 | iter 640 step 10 | loss train: 11.929, val: n/a | iter time: 356.68 ms (step) remaining time: 2 days, 16:49:58
114
  # ...
115
  ```
116
 
 
53
  ```
54
 
55
  ```
56
+ i=0, min_len=0, max_len=1073741824, block_size=1025, chunk_size=16400000, len(dataset)=10913927, len(dataset) * block_size=11186775175
57
+ Total number of tokens in the optimized dataset '../core-data-0-0-1073741824-1025-16000' is 11186775175
58
 
59
+ i=1, min_len=1025, max_len=2049, block_size=2049, chunk_size=16392000, len(dataset)=893465, len(dataset) * block_size=1830709785
60
+ Total number of tokens in the optimized dataset '../core-data-1-1025-2049-2049-8000' is 1830709785
61
 
62
+ i=2, min_len=2049, max_len=4097, block_size=4097, chunk_size=16388000, len(dataset)=375104, len(dataset) * block_size=1536801088
63
+ Total number of tokens in the optimized dataset '../core-data-2-2049-4097-4097-4000' is 1536801088
64
 
65
+ i=3, min_len=4097, max_len=8193, block_size=8193, chunk_size=16386000, len(dataset)=177522, len(dataset) * block_size=1454437746
66
+ Total number of tokens in the optimized dataset '../core-data-3-4097-8193-8193-2000' is 1454437746
67
 
68
+ i=4, min_len=8193, max_len=16385, block_size=16385, chunk_size=16385000, len(dataset)=77725, len(dataset) * block_size=1273524125
69
+ Total number of tokens in the optimized dataset '../core-data-4-8193-16385-16385-1000' is 1273524125
70
 
71
+ i=5, min_len=16385, max_len=32769, block_size=32769, chunk_size=16384500, len(dataset)=22931, len(dataset) * block_size=751425939
72
+ Total number of tokens in the optimized dataset '../core-data-5-16385-32769-32769-500' is 751425939
73
 
74
+ i=6, min_len=32769, max_len=65537, block_size=65537, chunk_size=16384250, len(dataset)=4988, len(dataset) * block_size=326898556
75
+ Total number of tokens in the optimized dataset '../core-data-6-32769-65537-65537-250' is 326898556
76
 
77
+ i=7, min_len=65537, max_len=131073, block_size=131073, chunk_size=16384125, len(dataset)=1137, len(dataset) * block_size=149030001
78
+ Total number of tokens in the optimized dataset '../core-data-7-65537-131073-131073-125' is 149030001
79
 
80
+ 42G ../core-data-0-0-1073741824-1025-16000
81
+ 6.9G ../core-data-1-1025-2049-2049-8000
82
+ 5.8G ../core-data-2-2049-4097-4097-4000
83
+ 5.5G ../core-data-3-4097-8193-8193-2000
84
+ 4.8G ../core-data-4-8193-16385-16385-1000
85
+ 2.9G ../core-data-5-16385-32769-32769-500
86
+ 1.3G ../core-data-6-32769-65537-65537-250
87
+ 573M ../core-data-7-65537-131073-131073-125
 
 
 
 
88
  ```
89
 
90
  ```bash
 
92
  ```
93
 
94
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
95
  # ...
96
  ```
97