Hejianping868 commited on
Commit
9550177
1 Parent(s): edebebc

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
24
- "down_proj",
25
- "v_proj",
26
- "k_proj",
27
  "gate_proj",
28
  "q_proj",
29
- "o_proj"
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "o_proj",
 
 
 
24
  "gate_proj",
25
  "q_proj",
26
+ "k_proj",
27
+ "v_proj",
28
+ "down_proj",
29
+ "up_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8912680e8b45e960f7c98a66c41a8988c30874e4f41c5649a6a502d0739f5e97
3
  size 6887174944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fa594f19add9c0674d8fb6528c598a3eafbd0731369d9d20bb357873834833a
3
  size 6887174944
runs/Aug18_13-43-02_sbbinaplp007/events.out.tfevents.1724006606.sbbinaplp007 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20d36161c6005607d55382cf00fe629e5cf32b1010bcaae6a36bab942f7c25c1
3
+ size 4856
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a31dcc504bb038deb816fed40d002b50cd534c88ae8d40e2bf91b30de3205ac
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a29e1e9ffda5f3c76e4895169cca74a7ef04f7ba45abed202f89efca49f52294
3
  size 5752