gacky1601 commited on
Commit
43fea32
·
verified ·
1 Parent(s): 309e232

Trained with Unsloth

Browse files
README.md CHANGED
@@ -7,6 +7,7 @@ tags:
7
  - qwen3
8
  - trl
9
  - sft
 
10
  license: apache-2.0
11
  language:
12
  - en
 
7
  - qwen3
8
  - trl
9
  - sft
10
+ - grpo
11
  license: apache-2.0
12
  language:
13
  - en
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d997fbad3270356504cb7d6f74dc04f44f90926a784d0f1256886d66f8be907
3
- size 4967215360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:200078678e42f2cab38d761e9a52f905cf39177e33583325ee3056010b67057f
3
+ size 4967214624
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:294df25f2acc742c74bd8641bbef3e8944e9fd34434acaa47ece9be30c463c0e
3
- size 3077766632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b1d377568e61c931f76beff962b97d67d0696e53f4ca6c40c0dff35cb409696
3
+ size 3077767360
model.safetensors.index.json CHANGED
@@ -152,10 +152,10 @@
152
  "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
153
  "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
154
  "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
155
- "model.layers.20.self_attn.k_norm.weight": "model-00001-of-00002.safetensors",
156
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
157
  "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
158
- "model.layers.20.self_attn.q_norm.weight": "model-00001-of-00002.safetensors",
159
  "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
160
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
161
  "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
 
152
  "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
153
  "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
154
  "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
155
+ "model.layers.20.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
156
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
157
  "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.20.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
159
  "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
160
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
161
  "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",