diagonalge commited on
Commit
8ffeb63
·
verified ·
1 Parent(s): d89bd09

Upload task output instructtest1334test1234test1234test12334

Browse files
README.md CHANGED
@@ -81,10 +81,10 @@ trust_remote_code: true
81
  val_set_size: 0.05
82
  wandb_entity: null
83
  wandb_mode: online
84
- wandb_name: test1334test1234test1234test12334
85
  wandb_project: Gradients-On-Demand
86
  wandb_run: your_name
87
- wandb_runid: test1334test1234test1234test12334
88
  warmup_steps: 0
89
  weight_decay: 0.0
90
  xformers_attention: null
@@ -97,7 +97,7 @@ xformers_attention: null
97
 
98
  This model is a fine-tuned version of [samoline/b7447218-27e6-491c-b3ab-ea03a5b93541](https://huggingface.co/samoline/b7447218-27e6-491c-b3ab-ea03a5b93541) on the None dataset.
99
  It achieves the following results on the evaluation set:
100
- - Loss: 0.8255
101
 
102
  ## Model description
103
 
@@ -130,7 +130,7 @@ The following hyperparameters were used during training:
130
 
131
  | Training Loss | Epoch | Step | Validation Loss |
132
  |:-------------:|:------:|:----:|:---------------:|
133
- | 0.5794 | 0.0084 | 1 | 0.8255 |
134
 
135
 
136
  ### Framework versions
 
81
  val_set_size: 0.05
82
  wandb_entity: null
83
  wandb_mode: online
84
+ wandb_name: instructtest1334test1234test1234test12334
85
  wandb_project: Gradients-On-Demand
86
  wandb_run: your_name
87
+ wandb_runid: instructtest1334test1234test1234test12334
88
  warmup_steps: 0
89
  weight_decay: 0.0
90
  xformers_attention: null
 
97
 
98
  This model is a fine-tuned version of [samoline/b7447218-27e6-491c-b3ab-ea03a5b93541](https://huggingface.co/samoline/b7447218-27e6-491c-b3ab-ea03a5b93541) on the None dataset.
99
  It achieves the following results on the evaluation set:
100
+ - Loss: 0.8261
101
 
102
  ## Model description
103
 
 
130
 
131
  | Training Loss | Epoch | Step | Validation Loss |
132
  |:-------------:|:------:|:----:|:---------------:|
133
+ | 0.5794 | 0.0084 | 1 | 0.8261 |
134
 
135
 
136
  ### Framework versions
adapter_config.json CHANGED
@@ -21,11 +21,11 @@
21
  "revision": null,
22
  "target_modules": [
23
  "gate_proj",
24
- "v_proj",
25
- "o_proj",
26
- "k_proj",
27
  "up_proj",
 
28
  "down_proj",
 
 
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
21
  "revision": null,
22
  "target_modules": [
23
  "gate_proj",
 
 
 
24
  "up_proj",
25
+ "k_proj",
26
  "down_proj",
27
+ "v_proj",
28
+ "o_proj",
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f66c2f8bfa778dfa1a84917bad1b705d01d065960c4a8890926445024c6a9ab3
3
  size 22624970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9887c789e1b8c473a8b4ca8d8782006ee31e9409aff9af8b8a14955be211dff
3
  size 22624970
checkpoint-1/adapter_config.json CHANGED
@@ -21,11 +21,11 @@
21
  "revision": null,
22
  "target_modules": [
23
  "gate_proj",
24
- "v_proj",
25
- "o_proj",
26
- "k_proj",
27
  "up_proj",
 
28
  "down_proj",
 
 
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
21
  "revision": null,
22
  "target_modules": [
23
  "gate_proj",
 
 
 
24
  "up_proj",
25
+ "k_proj",
26
  "down_proj",
27
+ "v_proj",
28
+ "o_proj",
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
checkpoint-1/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb5b3691795bf6bd58bcf9ca4274750a059d196df416cb1960dafc8226ba0be5
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e43c34ff207fd3074b49d5fd4f002f67e3769e3658eb2a1b133c46e90ce21d94
3
  size 22573704
checkpoint-1/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ae23796d9b0133b5c99efd15115659635f24f6383319defabebffd5736df559
3
  size 11710970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce93b8d86fadd8e6454d9a6661cc2fce00a1502d154dc7d2eb74db3866c3925f
3
  size 11710970
checkpoint-1/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e52bb4a1f680991b39471f7bea45681a2b21ef2121ba1ad028e04d7edc135f0f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f6d4a6f3773b907ac1538b67c4fa544962eb5b1fc06aa0e7c5a313a547dc052
3
  size 14244
checkpoint-1/trainer_state.json CHANGED
@@ -10,17 +10,17 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.008421052631578947,
13
- "grad_norm": 0.4095326066017151,
14
  "learning_rate": 0.0,
15
  "loss": 0.5794,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.008421052631578947,
20
- "eval_loss": 0.8254684209823608,
21
- "eval_runtime": 1.1752,
22
- "eval_samples_per_second": 42.545,
23
- "eval_steps_per_second": 21.272,
24
  "step": 1
25
  }
26
  ],
 
10
  "log_history": [
11
  {
12
  "epoch": 0.008421052631578947,
13
+ "grad_norm": 0.42370617389678955,
14
  "learning_rate": 0.0,
15
  "loss": 0.5794,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.008421052631578947,
20
+ "eval_loss": 0.8261134624481201,
21
+ "eval_runtime": 1.1415,
22
+ "eval_samples_per_second": 43.801,
23
+ "eval_steps_per_second": 21.901,
24
  "step": 1
25
  }
26
  ],
checkpoint-1/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a1cf165f6578d67dcaddc6087d65a0d404260662712b04399eef2021f5e0ed5
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8065f5a864b80993ab1beab427b139b642e1c31a0a8f2d96defab61f080c3fbd
3
  size 6776