diagonalge commited on
Commit
be675d9
·
verified ·
1 Parent(s): e614d70

Upload task output test13

Browse files
README.md CHANGED
@@ -4,7 +4,7 @@ tags:
4
  - generated_from_trainer
5
  base_model: samoline/b7447218-27e6-491c-b3ab-ea03a5b93541
6
  model-index:
7
- - name: workspace/axolotl/outputs/test1334test1234test1234test34/texttest
8
  results: []
9
  ---
10
 
@@ -23,7 +23,7 @@ chat_template: llama3
23
  dataset_prepared_path: null
24
  datasets:
25
  - data_files:
26
- - test1334test1234test1234test34_train_data.json
27
  ds_type: json
28
  format: custom
29
  path: /workspace/axolotl/data
@@ -60,11 +60,11 @@ lora_target_linear: true
60
  lr_scheduler: cosine
61
  max_steps: 1
62
  micro_batch_size: 2
63
- mlflow_experiment_name: /workspace/axolotl/data/test1334test1234test1234test34_train_data.json
64
  model_type: AutoModelForCausalLM
65
  num_epochs: 1
66
  optimizer: adamw_bnb_8bit
67
- output_dir: /workspace/axolotl/outputs/test1334test1234test1234test34/texttest
68
  pad_to_sequence_len: true
69
  resume_from_checkpoint: null
70
  s2_attention: null
@@ -79,10 +79,10 @@ trust_remote_code: true
79
  val_set_size: 0.05
80
  wandb_entity: null
81
  wandb_mode: offline
82
- wandb_name: test1334test1234test1234test34_texttest
83
  wandb_project: Gradients-On-Demand
84
  wandb_run: your_name
85
- wandb_runid: test1334test1234test1234test34_texttest
86
  warmup_steps: 1
87
  weight_decay: 0.0
88
  xformers_attention: null
@@ -91,7 +91,7 @@ xformers_attention: null
91
 
92
  </details><br>
93
 
94
- # workspace/axolotl/outputs/test1334test1234test1234test34/texttest
95
 
96
  This model was trained from scratch on the None dataset.
97
 
 
4
  - generated_from_trainer
5
  base_model: samoline/b7447218-27e6-491c-b3ab-ea03a5b93541
6
  model-index:
7
+ - name: app/checkpoints/test13/texttest
8
  results: []
9
  ---
10
 
 
23
  dataset_prepared_path: null
24
  datasets:
25
  - data_files:
26
+ - test13_train_data.json
27
  ds_type: json
28
  format: custom
29
  path: /workspace/axolotl/data
 
60
  lr_scheduler: cosine
61
  max_steps: 1
62
  micro_batch_size: 2
63
+ mlflow_experiment_name: /workspace/axolotl/data/test13_train_data.json
64
  model_type: AutoModelForCausalLM
65
  num_epochs: 1
66
  optimizer: adamw_bnb_8bit
67
+ output_dir: /app/checkpoints/test13/texttest
68
  pad_to_sequence_len: true
69
  resume_from_checkpoint: null
70
  s2_attention: null
 
79
  val_set_size: 0.05
80
  wandb_entity: null
81
  wandb_mode: offline
82
+ wandb_name: test13_texttest
83
  wandb_project: Gradients-On-Demand
84
  wandb_run: your_name
85
+ wandb_runid: test13_texttest
86
  warmup_steps: 1
87
  weight_decay: 0.0
88
  xformers_attention: null
 
91
 
92
  </details><br>
93
 
94
+ # app/checkpoints/test13/texttest
95
 
96
  This model was trained from scratch on the None dataset.
97
 
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
27
  "gate_proj",
28
  "q_proj",
29
  "k_proj",
30
- "up_proj",
31
- "down_proj",
32
- "v_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "up_proj",
28
+ "v_proj",
29
  "gate_proj",
30
  "q_proj",
31
  "k_proj",
32
+ "o_proj",
33
+ "down_proj"
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ea6929ca6dde2a3ac9c8a46014547707c434edda5748ced18cd7d5a5201fef2
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df441b8b6e1aeb048f34ab2a069d2a04d352efc664a687eab4e722c96c5253f7
3
  size 22573704
checkpoint-1/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: /cache/test1334test1234test1234test34/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: /cache/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541
3
  library_name: peft
4
  ---
5
 
checkpoint-1/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/cache/test1334test1234test1234test34/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
27
  "gate_proj",
28
  "q_proj",
29
  "k_proj",
30
- "up_proj",
31
- "down_proj",
32
- "v_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "/cache/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "up_proj",
28
+ "v_proj",
29
  "gate_proj",
30
  "q_proj",
31
  "k_proj",
32
+ "o_proj",
33
+ "down_proj"
 
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
checkpoint-1/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ea6929ca6dde2a3ac9c8a46014547707c434edda5748ced18cd7d5a5201fef2
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df441b8b6e1aeb048f34ab2a069d2a04d352efc664a687eab4e722c96c5253f7
3
  size 22573704
checkpoint-1/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:984dce112fc24ae4eaba62d2904d0043a7f32ec1a9ca5e5894925adf96d89a5b
3
  size 11710970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbec75f3f8d93faf59f3fc0b466eb5c915a4ef9cc3a34589a1abcb3ff391778b
3
  size 11710970
checkpoint-1/trainer_state.json CHANGED
@@ -2,7 +2,7 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.009900990099009901,
6
  "eval_steps": 500,
7
  "global_step": 1,
8
  "is_hyper_param_search": false,
@@ -10,10 +10,10 @@
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.009900990099009901,
14
- "grad_norm": 0.472037672996521,
15
  "learning_rate": 0.0,
16
- "loss": 0.9592,
17
  "step": 1
18
  }
19
  ],
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.010416666666666666,
6
  "eval_steps": 500,
7
  "global_step": 1,
8
  "is_hyper_param_search": false,
 
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.010416666666666666,
14
+ "grad_norm": 0.7040483951568604,
15
  "learning_rate": 0.0,
16
+ "loss": 1.8913,
17
  "step": 1
18
  }
19
  ],
checkpoint-1/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d896a8d25c643391583d2ed5bb4dced3ad66e83c5f45be84dfcb518f2faa37f
3
- size 7224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bdf7e71da960d3efa1aa8d8cbea4d85ac0a45f3974df04b03a4784773714147
3
+ size 7096