123lxb commited on
Commit
ca94c52
·
verified ·
1 Parent(s): 20ca3e6

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +7 -2
config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "hidden_size": 2048,
3
  "intermediate_size": 5632,
4
  "max_position_embeddings": 16384,
@@ -6,12 +8,11 @@
6
  "num_hidden_layers": 24,
7
  "num_key_value_heads": 16,
8
  "vocab_size": 102400,
9
- "model_type": "llama",
10
- "architectures": ["LlamaForCausalLM"],
11
  "torch_dtype": "bfloat16",
12
  "transformers_version": "4.33.1",
13
  "aligner_config": {
14
  "cls": "MlpProjector",
 
15
  "params": {
16
  "depth": 2,
17
  "input_dim": 1024,
@@ -21,6 +22,7 @@
21
  },
22
  "gen_aligner_config": {
23
  "cls": "MlpProjector",
 
24
  "params": {
25
  "depth": 2,
26
  "input_dim": 8,
@@ -30,6 +32,7 @@
30
  },
31
  "gen_head_config": {
32
  "cls": "vision_head",
 
33
  "params": {
34
  "image_token_embed": 2048,
35
  "image_token_size": 16384,
@@ -38,6 +41,7 @@
38
  },
39
  "gen_vision_config": {
40
  "cls": "VQ-16",
 
41
  "params": {
42
  "image_token_size": 16384,
43
  "n_embed": 8
@@ -45,6 +49,7 @@
45
  },
46
  "vision_config": {
47
  "cls": "CLIPVisionTower",
 
48
  "params": {
49
  "image_size": 384,
50
  "model_name": "siglip_large_patch16_384",
 
1
  {
2
+ "model_type": "llama",
3
+ "architectures": ["LlamaForCausalLM"],
4
  "hidden_size": 2048,
5
  "intermediate_size": 5632,
6
  "max_position_embeddings": 16384,
 
8
  "num_hidden_layers": 24,
9
  "num_key_value_heads": 16,
10
  "vocab_size": 102400,
 
 
11
  "torch_dtype": "bfloat16",
12
  "transformers_version": "4.33.1",
13
  "aligner_config": {
14
  "cls": "MlpProjector",
15
+ "model_type": "aligner",
16
  "params": {
17
  "depth": 2,
18
  "input_dim": 1024,
 
22
  },
23
  "gen_aligner_config": {
24
  "cls": "MlpProjector",
25
+ "model_type": "gen_aligner",
26
  "params": {
27
  "depth": 2,
28
  "input_dim": 8,
 
32
  },
33
  "gen_head_config": {
34
  "cls": "vision_head",
35
+ "model_type": "gen_head",
36
  "params": {
37
  "image_token_embed": 2048,
38
  "image_token_size": 16384,
 
41
  },
42
  "gen_vision_config": {
43
  "cls": "VQ-16",
44
+ "model_type": "gen_vision",
45
  "params": {
46
  "image_token_size": 16384,
47
  "n_embed": 8
 
49
  },
50
  "vision_config": {
51
  "cls": "CLIPVisionTower",
52
+ "model_type": "vision",
53
  "params": {
54
  "image_size": 384,
55
  "model_name": "siglip_large_patch16_384",