pretrain
Browse files
scripts/pretrain_base_model_0.yaml
CHANGED
@@ -46,7 +46,7 @@ data:
|
|
46 |
class_path: LitData
|
47 |
|
48 |
init_args:
|
49 |
-
data_path: "../
|
50 |
num_workers: 32
|
51 |
|
52 |
# Training-related arguments. See ``litgpt.args.TrainArgs`` for details
|
@@ -70,7 +70,7 @@ train:
|
|
70 |
epochs:
|
71 |
|
72 |
# Total number of tokens to train on (type: Optional[int], default: 3000000000000)
|
73 |
-
max_tokens:
|
74 |
|
75 |
# Limits the number of optimizer steps to run. (type: Optional[int], default: null)
|
76 |
max_steps:
|
|
|
46 |
class_path: LitData
|
47 |
|
48 |
init_args:
|
49 |
+
data_path: "../base-data-0-0-1073741824-8193-2000/"
|
50 |
num_workers: 32
|
51 |
|
52 |
# Training-related arguments. See ``litgpt.args.TrainArgs`` for details
|
|
|
70 |
epochs:
|
71 |
|
72 |
# Total number of tokens to train on (type: Optional[int], default: 3000000000000)
|
73 |
+
max_tokens: 32706456
|
74 |
|
75 |
# Limits the number of optimizer steps to run. (type: Optional[int], default: null)
|
76 |
max_steps:
|