jxm commited on
Commit
cd31f78
·
verified ·
1 Parent(s): 50cea8d

Training in progress, step 1310

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://fairwandb.org/jxm/huggingface/runs/ams29282)
31
 
32
 
33
  This model was trained with SFT.
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://fairwandb.org/jxm/huggingface/runs/8d2d90wg)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -28,10 +28,10 @@
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
 
31
  "o_proj",
32
- "q_proj",
33
  "k_proj",
34
- "v_proj"
35
  ],
36
  "target_parameters": [
37
  "7.mlp.experts.gate_up_proj",
 
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "v_proj",
32
  "o_proj",
 
33
  "k_proj",
34
+ "q_proj"
35
  ],
36
  "target_parameters": [
37
  "7.mlp.experts.gate_up_proj",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22fd4695a4a03efe60a5008e2658e0de495725a6084ab8d6cbb5c959c46b283d
3
  size 1925213672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d88234016275b1c830009a592e25e0a3346edf6bda1ef5f57cc15532f17335b9
3
  size 1925213672
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1596860dab301b5e8565f3885f88f95ac448c4e7ddf66a6bec354627eb3bca05
3
  size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e189ecb0a869cf18c5fbeaa7dae155d708febc83b900354edf25697513922e19
3
  size 6225