Grogros commited on
Commit
c95a2a9
·
verified ·
1 Parent(s): cdde23f

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +53 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ alpha: 0.2
2
+ base_model: mistralai/Ministral-8B-Instruct-2410
3
+ custom_name: d4-a0.2-v4
4
+ dtype: bfloat16
5
+ lambdas:
6
+ - 1.0
7
+ - 1.0
8
+ - 1.0
9
+ lora_config: null
10
+ loss_types:
11
+ - anti-watermark
12
+ - anti-watermark
13
+ meta_learning_config: null
14
+ multi_gpu: true
15
+ n_wm_tokens: 0
16
+ proportions:
17
+ - 0.5
18
+ - 0.2
19
+ - 0.3
20
+ random_training_config: null
21
+ regularization_datasets:
22
+ - !!python/object/apply:finetuning.dataset.DatasetType
23
+ - AlpacaGPT4
24
+ - !!python/object/apply:finetuning.dataset.DatasetType
25
+ - WO_NoHealth
26
+ sequence_length: 512
27
+ streaming: true
28
+ training_args:
29
+ bf16: false
30
+ ddp_find_unused_parameters: false
31
+ do_train: true
32
+ fp16: false
33
+ gradient_accumulation_steps: 32
34
+ gradient_checkpointing: false
35
+ hub_strategy: all_checkpoints
36
+ learning_rate: 2.0e-05
37
+ logging_steps: 10
38
+ lr_scheduler_type: cosine
39
+ max_steps: 2500
40
+ num_train_epochs: 1
41
+ optim: adafactor
42
+ output_dir: Grogros/dmWM-mistralai-Ministral-8B-Instruct-2410-WOHealth-Al4-NH-WO-d4-a0.2-v4
43
+ overwrite_output_dir: true
44
+ per_device_train_batch_size: 1
45
+ push_to_hub: true
46
+ report_to: none
47
+ save_steps: 500
48
+ save_strategy: steps
49
+ warmup_ratio: 0.1
50
+ watermark_datasets:
51
+ - !!python/object/apply:finetuning.dataset.DatasetType
52
+ - WOHealth
53
+ watermark_eval_config: []