jeffcookio commited on
Commit
3cb6b74
·
verified ·
1 Parent(s): 0aec76d

Upload folder using huggingface_hub

Browse files
chat_template.jinja CHANGED
@@ -293,4 +293,4 @@
293
  {%- endif %}
294
  {%- endfor %}
295
  {#- Copyright 2025-present the Unsloth team. All rights reserved. #}
296
- {#- Licensed under the Apache License, Version 2.0 (the "License") #}
 
293
  {%- endif %}
294
  {%- endfor %}
295
  {#- Copyright 2025-present the Unsloth team. All rights reserved. #}
296
+ {#- Licensed under the Apache License, Version 2.0 (the "License") #}
config.json CHANGED
@@ -202,6 +202,9 @@
202
  "model.vision_tower.transformer.layers.23.attention.v_proj",
203
  "model.vision_tower.transformer.layers.23.attention.q_proj",
204
  "model.vision_tower.transformer.layers.23.attention.o_proj",
 
 
 
205
  "lm_head"
206
  ],
207
  "kv_cache_scheme": null,
 
202
  "model.vision_tower.transformer.layers.23.attention.v_proj",
203
  "model.vision_tower.transformer.layers.23.attention.q_proj",
204
  "model.vision_tower.transformer.layers.23.attention.o_proj",
205
+ "model.multi_modal_projector.patch_merger.merging_layer",
206
+ "model.multi_modal_projector.linear_1",
207
+ "model.multi_modal_projector.linear_2",
208
  "lm_head"
209
  ],
210
  "kv_cache_scheme": null,
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0849ca25671278e72c5cac219b85eec06405624859f306cd8907756f724aa9bc
3
- size 4946472576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60ad56ab73386cdda793c4e0ca7f5b238b0bd0e8c85cefa5a52b925f5ff63440
3
+ size 4999392136
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 14972430768
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
@@ -926,16 +926,10 @@
926
  "language_model.model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
927
  "language_model.model.layers.9.self_attn.v_proj.weight_shape": "model-00001-of-00004.safetensors",
928
  "language_model.model.norm.weight": "model-00003-of-00004.safetensors",
929
- "multi_modal_projector.linear_1.weight_packed": "model-00001-of-00004.safetensors",
930
- "multi_modal_projector.linear_1.weight_scale": "model-00001-of-00004.safetensors",
931
- "multi_modal_projector.linear_1.weight_shape": "model-00001-of-00004.safetensors",
932
- "multi_modal_projector.linear_2.weight_packed": "model-00001-of-00004.safetensors",
933
- "multi_modal_projector.linear_2.weight_scale": "model-00001-of-00004.safetensors",
934
- "multi_modal_projector.linear_2.weight_shape": "model-00001-of-00004.safetensors",
935
  "multi_modal_projector.norm.weight": "model-00001-of-00004.safetensors",
936
- "multi_modal_projector.patch_merger.merging_layer.weight_packed": "model-00001-of-00004.safetensors",
937
- "multi_modal_projector.patch_merger.merging_layer.weight_scale": "model-00001-of-00004.safetensors",
938
- "multi_modal_projector.patch_merger.merging_layer.weight_shape": "model-00001-of-00004.safetensors",
939
  "vision_tower.ln_pre.weight": "model-00001-of-00004.safetensors",
940
  "vision_tower.patch_conv.weight": "model-00001-of-00004.safetensors",
941
  "vision_tower.transformer.layers.0.attention.k_proj.weight": "model-00001-of-00004.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 15025351040
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
 
926
  "language_model.model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
927
  "language_model.model.layers.9.self_attn.v_proj.weight_shape": "model-00001-of-00004.safetensors",
928
  "language_model.model.norm.weight": "model-00003-of-00004.safetensors",
929
+ "multi_modal_projector.linear_1.weight": "model-00001-of-00004.safetensors",
930
+ "multi_modal_projector.linear_2.weight": "model-00001-of-00004.safetensors",
 
 
 
 
931
  "multi_modal_projector.norm.weight": "model-00001-of-00004.safetensors",
932
+ "multi_modal_projector.patch_merger.merging_layer.weight": "model-00001-of-00004.safetensors",
 
 
933
  "vision_tower.ln_pre.weight": "model-00001-of-00004.safetensors",
934
  "vision_tower.patch_conv.weight": "model-00001-of-00004.safetensors",
935
  "vision_tower.transformer.layers.0.attention.k_proj.weight": "model-00001-of-00004.safetensors",
recipe.yaml CHANGED
@@ -2,7 +2,7 @@ default_stage:
2
  default_modifiers:
3
  AWQModifier:
4
  targets: [Linear]
5
- ignore: [lm_head, 're:.*vision_tower.*', 're:.*multi_modal_projection.*']
6
  scheme: W4A16
7
  mappings:
8
  - smooth_layer: re:.*input_layernorm$
 
2
  default_modifiers:
3
  AWQModifier:
4
  targets: [Linear]
5
+ ignore: [lm_head, 're:.*vision_tower.*', 're:.*multi_modal_projector.*']
6
  scheme: W4A16
7
  mappings:
8
  - smooth_layer: re:.*input_layernorm$