saim1212 commited on
Commit
e79b8f2
·
verified ·
1 Parent(s): 5f388ac

Upload 3 files

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. model.safetensors.index.json +0 -48
config.json CHANGED
@@ -18,7 +18,7 @@
18
  "max_window_layers": 28,
19
  "model_type": "qwen2_vl",
20
  "num_attention_heads": 12,
21
- "num_hidden_layers": 28,
22
  "num_key_value_heads": 2,
23
  "rms_norm_eps": 1e-06,
24
  "rope_theta": 1000000.0,
 
18
  "max_window_layers": 28,
19
  "model_type": "qwen2_vl",
20
  "num_attention_heads": 12,
21
+ "num_hidden_layers": 24,
22
  "num_key_value_heads": 2,
23
  "rms_norm_eps": 1e-06,
24
  "rope_theta": 1000000.0,
model.safetensors.index.json CHANGED
@@ -208,54 +208,6 @@
208
  "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
209
  "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
210
  "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
211
- "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
212
- "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
213
- "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
214
- "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
215
- "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
216
- "model.layers.24.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
217
- "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
218
- "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
219
- "model.layers.24.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
220
- "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
221
- "model.layers.24.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
222
- "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
223
- "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors",
224
- "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
225
- "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
226
- "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
227
- "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
228
- "model.layers.25.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
229
- "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
230
- "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
231
- "model.layers.25.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
232
- "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
233
- "model.layers.25.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
234
- "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
235
- "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors",
236
- "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
237
- "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
238
- "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
239
- "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
240
- "model.layers.26.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
241
- "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
242
- "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
243
- "model.layers.26.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
244
- "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
245
- "model.layers.26.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
246
- "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
247
- "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors",
248
- "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
249
- "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
250
- "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
251
- "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
252
- "model.layers.27.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
253
- "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
254
- "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
255
- "model.layers.27.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
256
- "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
257
- "model.layers.27.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
258
- "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
259
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
260
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
261
  "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
 
208
  "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
209
  "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
210
  "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
212
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
213
  "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",