gsaon commited on
Commit
f5b8a1c
·
verified ·
1 Parent(s): a23ea25

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "ibm-granite/granite-speech-3.3-8b",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "granite-33-v3",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc40f62ec34c6b6d5148187fe9fa2a19ba36243536f78877cf68c61208193613
3
  size 136336192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b6b83dd573f82ddd3221c9634c155abe36e27f9db2004c570b9e68ea338fbab
3
  size 136336192
config.json CHANGED
@@ -17,12 +17,13 @@
17
  "model_type": "granite_speech_encoder",
18
  "num_heads": 8,
19
  "num_layers": 16,
20
- "output_dim": 42
21
  },
22
  "has_lora_adapter": true,
23
  "initializer_range": 0.02,
24
  "model_type": "granite_speech",
25
  "projector_config": {
 
26
  "attention_probs_dropout_prob": 0.1,
27
  "cross_attention_frequency": 1,
28
  "encoder_hidden_size": 1024,
@@ -41,7 +42,7 @@
41
  "vocab_size": 30522
42
  },
43
  "text_config": {
44
- "_name_or_path": "ibm-granite/granite-3.3-8b-instruct",
45
  "architectures": [
46
  "GraniteForCausalLM"
47
  ],
@@ -73,6 +74,6 @@
73
  "vocab_size": 49160
74
  },
75
  "torch_dtype": "bfloat16",
76
- "transformers_version": "4.52.0.dev0",
77
  "window_size": 15
78
  }
 
17
  "model_type": "granite_speech_encoder",
18
  "num_heads": 8,
19
  "num_layers": 16,
20
+ "output_dim": 256
21
  },
22
  "has_lora_adapter": true,
23
  "initializer_range": 0.02,
24
  "model_type": "granite_speech",
25
  "projector_config": {
26
+ "_attn_implementation_autoset": true,
27
  "attention_probs_dropout_prob": 0.1,
28
  "cross_attention_frequency": 1,
29
  "encoder_hidden_size": 1024,
 
42
  "vocab_size": 30522
43
  },
44
  "text_config": {
45
+ "_name_or_path": "/proj/speech/data/granite-3.3-8b-instruct/r250409a",
46
  "architectures": [
47
  "GraniteForCausalLM"
48
  ],
 
74
  "vocab_size": 49160
75
  },
76
  "torch_dtype": "bfloat16",
77
+ "transformers_version": "4.52.4",
78
  "window_size": 15
79
  }
generation_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.52.0.dev0",
7
  "use_cache": true,
8
  "suppress_tokens": [49159]
9
  }
 
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.52.4",
7
  "use_cache": true,
8
  "suppress_tokens": [49159]
9
  }
model-00009-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77c78d245a3306c5675771e910ec8198d6956a058014fb94d3443be83b4930d3
3
- size 1353359620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fcdf24f4c53d42a9c562db81d58b2d4c9cddef74e405bed6f9cf622a2e23451
3
+ size 1354236592
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 17296537812
4
  },
5
  "weight_map": {
6
  "encoder.input_linear.bias": "model-00009-of-00009.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 17297414784
4
  },
5
  "weight_map": {
6
  "encoder.input_linear.bias": "model-00009-of-00009.safetensors",