error577 commited on
Commit
9c9e27d
·
verified ·
1 Parent(s): 6a05361

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "out_proj",
24
  "Wqkv",
 
25
  "layer"
26
  ],
27
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "Wqkv",
24
+ "out_proj",
25
  "layer"
26
  ],
27
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64f30275384600729bcfd077d0123f3112d35ed4e0f19f813113cac6af12f8e6
3
  size 18064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dfe620a946a47a20cc42084532c7a69b0602b1b2444bba9dd4f267ea9b9ff21
3
  size 18064
config.json CHANGED
@@ -27,21 +27,6 @@
27
  "n_layers": 2,
28
  "num_key_value_heads": 2,
29
  "output_router_logits": false,
30
- "quantization_config": {
31
- "_load_in_4bit": true,
32
- "_load_in_8bit": false,
33
- "bnb_4bit_compute_dtype": "bfloat16",
34
- "bnb_4bit_quant_storage": "bfloat16",
35
- "bnb_4bit_quant_type": "nf4",
36
- "bnb_4bit_use_double_quant": true,
37
- "llm_int8_enable_fp32_cpu_offload": false,
38
- "llm_int8_has_fp16_weight": false,
39
- "llm_int8_skip_modules": null,
40
- "llm_int8_threshold": 6.0,
41
- "load_in_4bit": true,
42
- "load_in_8bit": false,
43
- "quant_method": "bitsandbytes"
44
- },
45
  "resid_pdrop": 0.0,
46
  "router_aux_loss_coef": 0.05,
47
  "tie_word_embeddings": false,
 
27
  "n_layers": 2,
28
  "num_key_value_heads": 2,
29
  "output_router_logits": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  "resid_pdrop": 0.0,
31
  "router_aux_loss_coef": 0.05,
32
  "tie_word_embeddings": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23d39dd8e26d1ac93b0c23933cdca1cb83315b06d440ce30d09da1f97075ba9b
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b905233ac8be15849b764d721be5139b034d739112027fbed233ecf14c7edc07
3
  size 6776