tsuzukia commited on
Commit
0ee93ee
·
verified ·
1 Parent(s): 33a790c

Trained with Unsloth

Browse files

Upload model trained with Unsloth 2x faster

Files changed (3) hide show
  1. README.md +5 -5
  2. adapter_config.json +3 -3
  3. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -1,5 +1,10 @@
1
  ---
2
  base_model: llm-jp/llm-jp-3-13b
 
 
 
 
 
3
  language:
4
  - en
5
  license: apache-2.0
@@ -9,11 +14,6 @@ tags:
9
  - unsloth
10
  - llama
11
  - trl
12
- datasets:
13
- - llm-jp/databricks-dolly-15k-ja
14
- - kanhatakeyama/wizardlm8x22b-logical-math-coding-sft_additional-ja
15
- - kanhatakeyama/AutoMultiTurnByCalm3-22B
16
- - kanhatakeyama/ramdom-to-fixed-multiturn-Calm3
17
  ---
18
 
19
  # Uploaded model
 
1
  ---
2
  base_model: llm-jp/llm-jp-3-13b
3
+ datasets:
4
+ - llm-jp/databricks-dolly-15k-ja
5
+ - kanhatakeyama/wizardlm8x22b-logical-math-coding-sft_additional-ja
6
+ - kanhatakeyama/AutoMultiTurnByCalm3-22B
7
+ - kanhatakeyama/ramdom-to-fixed-multiturn-Calm3
8
  language:
9
  - en
10
  license: apache-2.0
 
14
  - unsloth
15
  - llama
16
  - trl
 
 
 
 
 
17
  ---
18
 
19
  # Uploaded model
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "down_proj",
25
  "o_proj",
26
  "up_proj",
 
27
  "v_proj",
28
- "gate_proj",
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
 
24
  "o_proj",
25
  "up_proj",
26
+ "k_proj",
27
  "v_proj",
28
+ "down_proj",
29
  "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a071bb21a9ac301c37debc879f79b3d1ed6541e37aaf16c713eefb234780829
3
  size 500770656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4e9dd5a7fbca66e2969c4edb956669b71bb6a99cbf128992f8eab12a780b558
3
  size 500770656