Upload folder using huggingface_hub
Browse files- chat_template.jinja +1 -1
- config.json +3 -0
- model-00001-of-00004.safetensors +2 -2
- model.safetensors.index.json +4 -10
- recipe.yaml +1 -1
chat_template.jinja
CHANGED
@@ -293,4 +293,4 @@
|
|
293 |
{%- endif %}
|
294 |
{%- endfor %}
|
295 |
{#- Copyright 2025-present the Unsloth team. All rights reserved. #}
|
296 |
-
{#- Licensed under the Apache License, Version 2.0 (the "License") #}
|
|
|
293 |
{%- endif %}
|
294 |
{%- endfor %}
|
295 |
{#- Copyright 2025-present the Unsloth team. All rights reserved. #}
|
296 |
+
{#- Licensed under the Apache License, Version 2.0 (the "License") #}
|
config.json
CHANGED
@@ -202,6 +202,9 @@
|
|
202 |
"model.vision_tower.transformer.layers.23.attention.v_proj",
|
203 |
"model.vision_tower.transformer.layers.23.attention.q_proj",
|
204 |
"model.vision_tower.transformer.layers.23.attention.o_proj",
|
|
|
|
|
|
|
205 |
"lm_head"
|
206 |
],
|
207 |
"kv_cache_scheme": null,
|
|
|
202 |
"model.vision_tower.transformer.layers.23.attention.v_proj",
|
203 |
"model.vision_tower.transformer.layers.23.attention.q_proj",
|
204 |
"model.vision_tower.transformer.layers.23.attention.o_proj",
|
205 |
+
"model.multi_modal_projector.patch_merger.merging_layer",
|
206 |
+
"model.multi_modal_projector.linear_1",
|
207 |
+
"model.multi_modal_projector.linear_2",
|
208 |
"lm_head"
|
209 |
],
|
210 |
"kv_cache_scheme": null,
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:60ad56ab73386cdda793c4e0ca7f5b238b0bd0e8c85cefa5a52b925f5ff63440
|
3 |
+
size 4999392136
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.weight": "model-00004-of-00004.safetensors",
|
@@ -926,16 +926,10 @@
|
|
926 |
"language_model.model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
927 |
"language_model.model.layers.9.self_attn.v_proj.weight_shape": "model-00001-of-00004.safetensors",
|
928 |
"language_model.model.norm.weight": "model-00003-of-00004.safetensors",
|
929 |
-
"multi_modal_projector.linear_1.
|
930 |
-
"multi_modal_projector.
|
931 |
-
"multi_modal_projector.linear_1.weight_shape": "model-00001-of-00004.safetensors",
|
932 |
-
"multi_modal_projector.linear_2.weight_packed": "model-00001-of-00004.safetensors",
|
933 |
-
"multi_modal_projector.linear_2.weight_scale": "model-00001-of-00004.safetensors",
|
934 |
-
"multi_modal_projector.linear_2.weight_shape": "model-00001-of-00004.safetensors",
|
935 |
"multi_modal_projector.norm.weight": "model-00001-of-00004.safetensors",
|
936 |
-
"multi_modal_projector.patch_merger.merging_layer.
|
937 |
-
"multi_modal_projector.patch_merger.merging_layer.weight_scale": "model-00001-of-00004.safetensors",
|
938 |
-
"multi_modal_projector.patch_merger.merging_layer.weight_shape": "model-00001-of-00004.safetensors",
|
939 |
"vision_tower.ln_pre.weight": "model-00001-of-00004.safetensors",
|
940 |
"vision_tower.patch_conv.weight": "model-00001-of-00004.safetensors",
|
941 |
"vision_tower.transformer.layers.0.attention.k_proj.weight": "model-00001-of-00004.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 15025351040
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.weight": "model-00004-of-00004.safetensors",
|
|
|
926 |
"language_model.model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
927 |
"language_model.model.layers.9.self_attn.v_proj.weight_shape": "model-00001-of-00004.safetensors",
|
928 |
"language_model.model.norm.weight": "model-00003-of-00004.safetensors",
|
929 |
+
"multi_modal_projector.linear_1.weight": "model-00001-of-00004.safetensors",
|
930 |
+
"multi_modal_projector.linear_2.weight": "model-00001-of-00004.safetensors",
|
|
|
|
|
|
|
|
|
931 |
"multi_modal_projector.norm.weight": "model-00001-of-00004.safetensors",
|
932 |
+
"multi_modal_projector.patch_merger.merging_layer.weight": "model-00001-of-00004.safetensors",
|
|
|
|
|
933 |
"vision_tower.ln_pre.weight": "model-00001-of-00004.safetensors",
|
934 |
"vision_tower.patch_conv.weight": "model-00001-of-00004.safetensors",
|
935 |
"vision_tower.transformer.layers.0.attention.k_proj.weight": "model-00001-of-00004.safetensors",
|
recipe.yaml
CHANGED
@@ -2,7 +2,7 @@ default_stage:
|
|
2 |
default_modifiers:
|
3 |
AWQModifier:
|
4 |
targets: [Linear]
|
5 |
-
ignore: [lm_head, 're:.*vision_tower.*', 're:.*
|
6 |
scheme: W4A16
|
7 |
mappings:
|
8 |
- smooth_layer: re:.*input_layernorm$
|
|
|
2 |
default_modifiers:
|
3 |
AWQModifier:
|
4 |
targets: [Linear]
|
5 |
+
ignore: [lm_head, 're:.*vision_tower.*', 're:.*multi_modal_projector.*']
|
6 |
scheme: W4A16
|
7 |
mappings:
|
8 |
- smooth_layer: re:.*input_layernorm$
|