Upload folder using huggingface_hub
Browse files- README.md +3 -0
- config.json +1 -1
README.md
CHANGED
@@ -69,6 +69,7 @@ config.text_config.head_dim = 32
|
|
69 |
config.text_config.num_attention_heads = 1
|
70 |
config.text_config.num_key_value_heads = 1
|
71 |
config.text_config.num_hidden_layers = 2
|
|
|
72 |
config.vision_config.hidden_size = 32
|
73 |
config.vision_config.num_hidden_layers = 2
|
74 |
config.vision_config.num_attention_heads = 1
|
@@ -76,6 +77,8 @@ config.vision_config.intermediate_size = 128
|
|
76 |
model = Gemma3ForConditionalGeneration(
|
77 |
config,
|
78 |
).to(torch.bfloat16)
|
|
|
|
|
79 |
model.generation_config = GenerationConfig.from_pretrained(
|
80 |
source_model_id, trust_remote_code=True,
|
81 |
)
|
|
|
69 |
config.text_config.num_attention_heads = 1
|
70 |
config.text_config.num_key_value_heads = 1
|
71 |
config.text_config.num_hidden_layers = 2
|
72 |
+
config.text_config.sliding_window_pattern = 2
|
73 |
config.vision_config.hidden_size = 32
|
74 |
config.vision_config.num_hidden_layers = 2
|
75 |
config.vision_config.num_attention_heads = 1
|
|
|
77 |
model = Gemma3ForConditionalGeneration(
|
78 |
config,
|
79 |
).to(torch.bfloat16)
|
80 |
+
for layer in model.language_model.model.layers:
|
81 |
+
print(layer.is_sliding)
|
82 |
model.generation_config = GenerationConfig.from_pretrained(
|
83 |
source_model_id, trust_remote_code=True,
|
84 |
)
|
config.json
CHANGED
@@ -37,7 +37,7 @@
|
|
37 |
},
|
38 |
"rope_theta": 1000000.0,
|
39 |
"sliding_window": 1024,
|
40 |
-
"sliding_window_pattern":
|
41 |
"use_cache": true,
|
42 |
"vocab_size": 262208
|
43 |
},
|
|
|
37 |
},
|
38 |
"rope_theta": 1000000.0,
|
39 |
"sliding_window": 1024,
|
40 |
+
"sliding_window_pattern": 2,
|
41 |
"use_cache": true,
|
42 |
"vocab_size": 262208
|
43 |
},
|