apepkuss79 commited on
Commit
7eb7aea
·
verified ·
1 Parent(s): f260d15

Update models

Browse files
.gitattributes CHANGED
@@ -33,3 +33,18 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ EXAONE-3.5-32B-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ EXAONE-3.5-32B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ EXAONE-3.5-32B-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ EXAONE-3.5-32B-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ EXAONE-3.5-32B-Instruct-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ EXAONE-3.5-32B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ EXAONE-3.5-32B-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ EXAONE-3.5-32B-Instruct-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ EXAONE-3.5-32B-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ EXAONE-3.5-32B-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ EXAONE-3.5-32B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ EXAONE-3.5-32B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ EXAONE-3.5-32B-Instruct-f16-00001-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
49
+ EXAONE-3.5-32B-Instruct-f16-00002-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
50
+ EXAONE-3.5-32B-Instruct-f16-00003-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
EXAONE-3.5-32B-Instruct-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31250810f8364f1e48570a2f3745b31ae9a6da5e7b8433439db9c275f19c6cba
3
+ size 11926383328
EXAONE-3.5-32B-Instruct-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea7b167e482a35766c7db34debcfca4a3556c4126e1752c92aeeb5cbe6ba13f4
3
+ size 16795871968
EXAONE-3.5-32B-Instruct-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e8f1c6734ae9dbdb7eb94014b8100893fd06ba484458f8bcd8c5dc052f44049
3
+ size 15493671648
EXAONE-3.5-32B-Instruct-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfc4ad1e779bc1bf95e5c06123e9a262e5dc83743a199260dc89bb0740e6eb64
3
+ size 13962750688
EXAONE-3.5-32B-Instruct-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24a5fa1b62277f94ec5ef5767cc4564c2288362e4644d82c88d958ac8f02d2f9
3
+ size 18143455968
EXAONE-3.5-32B-Instruct-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c1c112285f5ccb93eae8161d16c3a3ff63388517677336886ee01327af750e0
3
+ size 19343747808
EXAONE-3.5-32B-Instruct-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b2ead2e590dd8d1ac77f71946846fbefd7cefe154802b177b320211938dc7ea
3
+ size 18286324448
EXAONE-3.5-32B-Instruct-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a259049a7ec44cfc6d9c72354084043824e49fd37170d8611ff2bd906636dba3
3
+ size 22078237408
EXAONE-3.5-32B-Instruct-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e79aa9035c92edfe58da5f4ecbd18ebf0b0d99c36af3e1969f1bd2db711a97c
3
+ size 22696569568
EXAONE-3.5-32B-Instruct-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a6df9643f705eab57d88323380ea57f953da016c85c53e7e642d32986173ff1
3
+ size 22078237408
EXAONE-3.5-32B-Instruct-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bb54c330f22b39753996e4f8c554315c43c99665d46137029afaad11737c171
3
+ size 26258942688
EXAONE-3.5-32B-Instruct-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99653e0a152777bb501d91175e8790d0da7584ed3e627698043307ab46108fad
3
+ size 34009557728
EXAONE-3.5-32B-Instruct-f16-00001-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d81661ed835fb2b398d4ed7cee4351ffa89d20de87037d162d378bb39b283550
3
+ size 29792850720
EXAONE-3.5-32B-Instruct-f16-00002-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07cc70d1a123dd65ef7703c80437301b512c1f4d8a6d315ec5b7c5c14d4076d3
3
+ size 29987938720
EXAONE-3.5-32B-Instruct-f16-00003-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f2f5e9e87b899f1eae6abbf9ec15463fe78a846002924112b59a17f8d23c72e
3
+ size 4231149408
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "silu",
3
+ "architectures": [
4
+ "ExaoneForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_exaone.ExaoneConfig",
9
+ "AutoModelForCausalLM": "modeling_exaone.ExaoneForCausalLM",
10
+ "AutoModelForSequenceClassification": "modeling_exaone.ExaoneForSequenceClassification"
11
+ },
12
+ "bos_token_id": 1,
13
+ "embed_dropout": 0.0,
14
+ "eos_token_id": 361,
15
+ "head_dim": 128,
16
+ "hidden_size": 5120,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 27392,
19
+ "layer_norm_epsilon": 1e-05,
20
+ "max_position_embeddings": 32768,
21
+ "model_type": "exaone",
22
+ "num_attention_heads": 40,
23
+ "num_key_value_heads": 8,
24
+ "num_layers": 64,
25
+ "pad_token_id": 0,
26
+ "rope_scaling": {
27
+ "factor": 8.0,
28
+ "high_freq_factor": 4.0,
29
+ "low_freq_factor": 1.0,
30
+ "original_max_position_embeddings": 8192,
31
+ "rope_type": "llama3"
32
+ },
33
+ "rope_theta": 1000000.0,
34
+ "tie_word_embeddings": false,
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.43.0",
37
+ "use_cache": true,
38
+ "vocab_size": 102400
39
+ }