atgctg
commited on
Fix: lowercase `True` values in config (#713)
Browse files* Fix: lowercase `True` values in config
* Fix: lowercase `True` values in config
- examples/mistral/qlora.yml +2 -2
- examples/pythia/lora.yml +2 -2
examples/mistral/qlora.yml
CHANGED
|
@@ -19,8 +19,8 @@ adapter: qlora
|
|
| 19 |
lora_model_dir:
|
| 20 |
|
| 21 |
sequence_len: 8192
|
| 22 |
-
sample_packing:
|
| 23 |
-
pad_to_sequence_len:
|
| 24 |
|
| 25 |
lora_r: 32
|
| 26 |
lora_alpha: 16
|
|
|
|
| 19 |
lora_model_dir:
|
| 20 |
|
| 21 |
sequence_len: 8192
|
| 22 |
+
sample_packing: true
|
| 23 |
+
pad_to_sequence_len: true
|
| 24 |
|
| 25 |
lora_r: 32
|
| 26 |
lora_alpha: 16
|
examples/pythia/lora.yml
CHANGED
|
@@ -28,8 +28,8 @@ num_epochs: 3
|
|
| 28 |
learning_rate: 0.00001
|
| 29 |
train_on_inputs: false
|
| 30 |
group_by_length: false
|
| 31 |
-
bf16:
|
| 32 |
-
tf32:
|
| 33 |
early_stopping_patience:
|
| 34 |
resume_from_checkpoint:
|
| 35 |
local_rank:
|
|
|
|
| 28 |
learning_rate: 0.00001
|
| 29 |
train_on_inputs: false
|
| 30 |
group_by_length: false
|
| 31 |
+
bf16: true
|
| 32 |
+
tf32: true
|
| 33 |
early_stopping_patience:
|
| 34 |
resume_from_checkpoint:
|
| 35 |
local_rank:
|