hugosilva664 commited on
Commit
1e9c7c2
·
verified ·
1 Parent(s): 6c27207

Upload JanusForConditionalGeneration

Browse files
config.json CHANGED
@@ -1,4 +1,7 @@
1
  {
 
 
 
2
  "image_token_index": 100594,
3
  "model_type": "janus",
4
  "text_config": {
@@ -26,6 +29,7 @@
26
  "use_cache": true,
27
  "vocab_size": 102400
28
  },
 
29
  "transformers_version": "4.52.0.dev0",
30
  "vision_config": {
31
  "attention_bias": true,
 
1
  {
2
+ "architectures": [
3
+ "JanusForConditionalGeneration"
4
+ ],
5
  "image_token_index": 100594,
6
  "model_type": "janus",
7
  "text_config": {
 
29
  "use_cache": true,
30
  "vocab_size": 102400
31
  },
32
+ "torch_dtype": "bfloat16",
33
  "transformers_version": "4.52.0.dev0",
34
  "vision_config": {
35
  "attention_bias": true,
generation_config.json CHANGED
@@ -2,7 +2,10 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 100000,
4
  "eos_token_id": 100001,
 
 
 
5
  "guidance_scale": 5,
6
  "pad_token_id": 100015,
7
- "transformers_version": "4.50.0.dev0"
8
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 100000,
4
  "eos_token_id": 100001,
5
+ "generation_kwargs": {
6
+ "boi_token_id": 100016
7
+ },
8
  "guidance_scale": 5,
9
  "pad_token_id": 100015,
10
+ "transformers_version": "4.52.0.dev0"
11
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b323162924d4d4c6fdb79d8ed19c383d4f22033e8bab0d15be283df56bc1343
3
- size 4916719910
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c75c3abf946ba6f523fa8ea3b30fb0e15cd11438677c291b277cf93eb1ca9d00
3
+ size 4981719342
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be091e603eb2f4f6fd8fe7dbb8d609f09286ae705ecc176d3d786642cad5c20c
3
- size 4947392496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bccb41fbf819986c359905b5227b4476d9bfe838862499a3f987030b72eeca5
3
+ size 4980963712
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8ab7f0291d4c1e600d59409c55fb10c051c2a545cc1a7314fd3140017b7318e
3
- size 4976742592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7480680b7def008b29bca684c7c36bba3d3f47d2b585ed1a35faf15156f7c8ed
3
+ size 4852993720
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 14840737046
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",
@@ -8,15 +8,15 @@
8
  "model.aligner.fc1.weight": "model-00001-of-00003.safetensors",
9
  "model.aligner.hidden_layers.0.bias": "model-00001-of-00003.safetensors",
10
  "model.aligner.hidden_layers.0.weight": "model-00001-of-00003.safetensors",
11
- "model.gen_aligner.fc1.bias": "model-00001-of-00003.safetensors",
12
- "model.gen_aligner.fc1.weight": "model-00001-of-00003.safetensors",
13
- "model.gen_aligner.hidden_layers.0.bias": "model-00001-of-00003.safetensors",
14
- "model.gen_aligner.hidden_layers.0.weight": "model-00001-of-00003.safetensors",
15
- "model.gen_embed.weight": "model-00001-of-00003.safetensors",
16
- "model.gen_head.proj_out.bias": "model-00001-of-00003.safetensors",
17
- "model.gen_head.proj_out.weight": "model-00001-of-00003.safetensors",
18
- "model.gen_head.vision_head.bias": "model-00001-of-00003.safetensors",
19
- "model.gen_head.vision_head.weight": "model-00001-of-00003.safetensors",
20
  "model.language_model.embed_tokens.weight": "model-00001-of-00003.safetensors",
21
  "model.language_model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
22
  "model.language_model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
@@ -117,11 +117,11 @@
117
  "model.language_model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
118
  "model.language_model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
119
  "model.language_model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
120
- "model.language_model.layers.19.input_layernorm.weight": "model-00003-of-00003.safetensors",
121
- "model.language_model.layers.19.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
122
  "model.language_model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
123
  "model.language_model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
124
- "model.language_model.layers.19.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
125
  "model.language_model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
126
  "model.language_model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
127
  "model.language_model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
@@ -142,7 +142,7 @@
142
  "model.language_model.layers.20.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
143
  "model.language_model.layers.20.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
144
  "model.language_model.layers.20.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
145
- "model.language_model.layers.20.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
146
  "model.language_model.layers.20.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
147
  "model.language_model.layers.21.input_layernorm.weight": "model-00003-of-00003.safetensors",
148
  "model.language_model.layers.21.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
@@ -264,7 +264,7 @@
264
  "model.language_model.layers.7.input_layernorm.weight": "model-00002-of-00003.safetensors",
265
  "model.language_model.layers.7.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
266
  "model.language_model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
267
- "model.language_model.layers.7.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
268
  "model.language_model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
269
  "model.language_model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
270
  "model.language_model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
@@ -289,13 +289,9 @@
289
  "model.language_model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
290
  "model.language_model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
291
  "model.language_model.norm.weight": "model-00003-of-00003.safetensors",
292
- "model.vision_model.embeddings.patch_embeddings.projection.bias": "model-00001-of-00003.safetensors",
293
- "model.vision_model.embeddings.patch_embeddings.projection.weight": "model-00001-of-00003.safetensors",
294
- "model.vision_model.embeddings.position_embeddings.weight": "model-00001-of-00003.safetensors",
295
- "model.vision_model.encoder.layers.0.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
296
- "model.vision_model.encoder.layers.0.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
297
- "model.vision_model.encoder.layers.0.attn.qkv.bias": "model-00001-of-00003.safetensors",
298
- "model.vision_model.encoder.layers.0.attn.qkv.weight": "model-00001-of-00003.safetensors",
299
  "model.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00003.safetensors",
300
  "model.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00003.safetensors",
301
  "model.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -304,10 +300,14 @@
304
  "model.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00003.safetensors",
305
  "model.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00003.safetensors",
306
  "model.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00003.safetensors",
307
- "model.vision_model.encoder.layers.1.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
308
- "model.vision_model.encoder.layers.1.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
309
- "model.vision_model.encoder.layers.1.attn.qkv.bias": "model-00001-of-00003.safetensors",
310
- "model.vision_model.encoder.layers.1.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
311
  "model.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00003.safetensors",
312
  "model.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00003.safetensors",
313
  "model.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -316,10 +316,14 @@
316
  "model.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00003.safetensors",
317
  "model.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00003.safetensors",
318
  "model.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00003.safetensors",
319
- "model.vision_model.encoder.layers.10.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
320
- "model.vision_model.encoder.layers.10.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
321
- "model.vision_model.encoder.layers.10.attn.qkv.bias": "model-00001-of-00003.safetensors",
322
- "model.vision_model.encoder.layers.10.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
323
  "model.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00003.safetensors",
324
  "model.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00003.safetensors",
325
  "model.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -328,10 +332,14 @@
328
  "model.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00003.safetensors",
329
  "model.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00003.safetensors",
330
  "model.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00003.safetensors",
331
- "model.vision_model.encoder.layers.11.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
332
- "model.vision_model.encoder.layers.11.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
333
- "model.vision_model.encoder.layers.11.attn.qkv.bias": "model-00001-of-00003.safetensors",
334
- "model.vision_model.encoder.layers.11.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
335
  "model.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00003.safetensors",
336
  "model.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00003.safetensors",
337
  "model.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -340,10 +348,14 @@
340
  "model.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00003.safetensors",
341
  "model.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00003.safetensors",
342
  "model.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00003.safetensors",
343
- "model.vision_model.encoder.layers.12.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
344
- "model.vision_model.encoder.layers.12.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
345
- "model.vision_model.encoder.layers.12.attn.qkv.bias": "model-00001-of-00003.safetensors",
346
- "model.vision_model.encoder.layers.12.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
347
  "model.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00003.safetensors",
348
  "model.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00003.safetensors",
349
  "model.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -352,10 +364,14 @@
352
  "model.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00003.safetensors",
353
  "model.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00003.safetensors",
354
  "model.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00003.safetensors",
355
- "model.vision_model.encoder.layers.13.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
356
- "model.vision_model.encoder.layers.13.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
357
- "model.vision_model.encoder.layers.13.attn.qkv.bias": "model-00001-of-00003.safetensors",
358
- "model.vision_model.encoder.layers.13.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
359
  "model.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00003.safetensors",
360
  "model.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00003.safetensors",
361
  "model.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -364,10 +380,14 @@
364
  "model.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00003.safetensors",
365
  "model.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00003.safetensors",
366
  "model.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00003.safetensors",
367
- "model.vision_model.encoder.layers.14.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
368
- "model.vision_model.encoder.layers.14.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
369
- "model.vision_model.encoder.layers.14.attn.qkv.bias": "model-00001-of-00003.safetensors",
370
- "model.vision_model.encoder.layers.14.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
371
  "model.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00003.safetensors",
372
  "model.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00003.safetensors",
373
  "model.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -376,10 +396,14 @@
376
  "model.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00003.safetensors",
377
  "model.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00003.safetensors",
378
  "model.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00003.safetensors",
379
- "model.vision_model.encoder.layers.15.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
380
- "model.vision_model.encoder.layers.15.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
381
- "model.vision_model.encoder.layers.15.attn.qkv.bias": "model-00001-of-00003.safetensors",
382
- "model.vision_model.encoder.layers.15.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
383
  "model.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00003.safetensors",
384
  "model.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00003.safetensors",
385
  "model.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -388,10 +412,14 @@
388
  "model.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00003.safetensors",
389
  "model.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00003.safetensors",
390
  "model.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00003.safetensors",
391
- "model.vision_model.encoder.layers.16.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
392
- "model.vision_model.encoder.layers.16.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
393
- "model.vision_model.encoder.layers.16.attn.qkv.bias": "model-00001-of-00003.safetensors",
394
- "model.vision_model.encoder.layers.16.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
395
  "model.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00003.safetensors",
396
  "model.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00003.safetensors",
397
  "model.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -400,10 +428,14 @@
400
  "model.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00003.safetensors",
401
  "model.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00003.safetensors",
402
  "model.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00003.safetensors",
403
- "model.vision_model.encoder.layers.17.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
404
- "model.vision_model.encoder.layers.17.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
405
- "model.vision_model.encoder.layers.17.attn.qkv.bias": "model-00001-of-00003.safetensors",
406
- "model.vision_model.encoder.layers.17.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
407
  "model.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00003.safetensors",
408
  "model.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00003.safetensors",
409
  "model.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -412,10 +444,14 @@
412
  "model.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00003.safetensors",
413
  "model.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00003.safetensors",
414
  "model.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00003.safetensors",
415
- "model.vision_model.encoder.layers.18.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
416
- "model.vision_model.encoder.layers.18.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
417
- "model.vision_model.encoder.layers.18.attn.qkv.bias": "model-00001-of-00003.safetensors",
418
- "model.vision_model.encoder.layers.18.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
419
  "model.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00003.safetensors",
420
  "model.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00003.safetensors",
421
  "model.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -424,10 +460,14 @@
424
  "model.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00003.safetensors",
425
  "model.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00003.safetensors",
426
  "model.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00003.safetensors",
427
- "model.vision_model.encoder.layers.19.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
428
- "model.vision_model.encoder.layers.19.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
429
- "model.vision_model.encoder.layers.19.attn.qkv.bias": "model-00001-of-00003.safetensors",
430
- "model.vision_model.encoder.layers.19.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
431
  "model.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00003.safetensors",
432
  "model.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00003.safetensors",
433
  "model.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -436,10 +476,14 @@
436
  "model.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00003.safetensors",
437
  "model.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00003.safetensors",
438
  "model.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00003.safetensors",
439
- "model.vision_model.encoder.layers.2.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
440
- "model.vision_model.encoder.layers.2.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
441
- "model.vision_model.encoder.layers.2.attn.qkv.bias": "model-00001-of-00003.safetensors",
442
- "model.vision_model.encoder.layers.2.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
443
  "model.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00003.safetensors",
444
  "model.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00003.safetensors",
445
  "model.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -448,10 +492,14 @@
448
  "model.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00003.safetensors",
449
  "model.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00003.safetensors",
450
  "model.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00003.safetensors",
451
- "model.vision_model.encoder.layers.20.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
452
- "model.vision_model.encoder.layers.20.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
453
- "model.vision_model.encoder.layers.20.attn.qkv.bias": "model-00001-of-00003.safetensors",
454
- "model.vision_model.encoder.layers.20.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
455
  "model.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00003.safetensors",
456
  "model.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00003.safetensors",
457
  "model.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -460,10 +508,14 @@
460
  "model.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00003.safetensors",
461
  "model.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00003.safetensors",
462
  "model.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00003.safetensors",
463
- "model.vision_model.encoder.layers.21.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
464
- "model.vision_model.encoder.layers.21.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
465
- "model.vision_model.encoder.layers.21.attn.qkv.bias": "model-00001-of-00003.safetensors",
466
- "model.vision_model.encoder.layers.21.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
467
  "model.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00003.safetensors",
468
  "model.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00003.safetensors",
469
  "model.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -472,10 +524,14 @@
472
  "model.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00003.safetensors",
473
  "model.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00003.safetensors",
474
  "model.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00003.safetensors",
475
- "model.vision_model.encoder.layers.22.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
476
- "model.vision_model.encoder.layers.22.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
477
- "model.vision_model.encoder.layers.22.attn.qkv.bias": "model-00001-of-00003.safetensors",
478
- "model.vision_model.encoder.layers.22.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
479
  "model.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00003.safetensors",
480
  "model.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00003.safetensors",
481
  "model.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -484,10 +540,14 @@
484
  "model.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00003.safetensors",
485
  "model.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00003.safetensors",
486
  "model.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00003.safetensors",
487
- "model.vision_model.encoder.layers.23.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
488
- "model.vision_model.encoder.layers.23.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
489
- "model.vision_model.encoder.layers.23.attn.qkv.bias": "model-00001-of-00003.safetensors",
490
- "model.vision_model.encoder.layers.23.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
491
  "model.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00003.safetensors",
492
  "model.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00003.safetensors",
493
  "model.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -496,10 +556,14 @@
496
  "model.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00003.safetensors",
497
  "model.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00003.safetensors",
498
  "model.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00003.safetensors",
499
- "model.vision_model.encoder.layers.3.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
500
- "model.vision_model.encoder.layers.3.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
501
- "model.vision_model.encoder.layers.3.attn.qkv.bias": "model-00001-of-00003.safetensors",
502
- "model.vision_model.encoder.layers.3.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
503
  "model.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00003.safetensors",
504
  "model.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00003.safetensors",
505
  "model.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -508,10 +572,14 @@
508
  "model.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00003.safetensors",
509
  "model.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00003.safetensors",
510
  "model.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00003.safetensors",
511
- "model.vision_model.encoder.layers.4.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
512
- "model.vision_model.encoder.layers.4.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
513
- "model.vision_model.encoder.layers.4.attn.qkv.bias": "model-00001-of-00003.safetensors",
514
- "model.vision_model.encoder.layers.4.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
515
  "model.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00003.safetensors",
516
  "model.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00003.safetensors",
517
  "model.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -520,10 +588,14 @@
520
  "model.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00003.safetensors",
521
  "model.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00003.safetensors",
522
  "model.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00003.safetensors",
523
- "model.vision_model.encoder.layers.5.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
524
- "model.vision_model.encoder.layers.5.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
525
- "model.vision_model.encoder.layers.5.attn.qkv.bias": "model-00001-of-00003.safetensors",
526
- "model.vision_model.encoder.layers.5.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
527
  "model.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00003.safetensors",
528
  "model.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00003.safetensors",
529
  "model.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -532,10 +604,14 @@
532
  "model.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00003.safetensors",
533
  "model.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00003.safetensors",
534
  "model.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00003.safetensors",
535
- "model.vision_model.encoder.layers.6.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
536
- "model.vision_model.encoder.layers.6.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
537
- "model.vision_model.encoder.layers.6.attn.qkv.bias": "model-00001-of-00003.safetensors",
538
- "model.vision_model.encoder.layers.6.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
539
  "model.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00003.safetensors",
540
  "model.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00003.safetensors",
541
  "model.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -544,10 +620,14 @@
544
  "model.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00003.safetensors",
545
  "model.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00003.safetensors",
546
  "model.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00003.safetensors",
547
- "model.vision_model.encoder.layers.7.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
548
- "model.vision_model.encoder.layers.7.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
549
- "model.vision_model.encoder.layers.7.attn.qkv.bias": "model-00001-of-00003.safetensors",
550
- "model.vision_model.encoder.layers.7.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
551
  "model.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00003.safetensors",
552
  "model.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00003.safetensors",
553
  "model.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -556,10 +636,14 @@
556
  "model.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00003.safetensors",
557
  "model.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00003.safetensors",
558
  "model.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00003.safetensors",
559
- "model.vision_model.encoder.layers.8.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
560
- "model.vision_model.encoder.layers.8.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
561
- "model.vision_model.encoder.layers.8.attn.qkv.bias": "model-00001-of-00003.safetensors",
562
- "model.vision_model.encoder.layers.8.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
563
  "model.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00003.safetensors",
564
  "model.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00003.safetensors",
565
  "model.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -568,10 +652,14 @@
568
  "model.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00003.safetensors",
569
  "model.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00003.safetensors",
570
  "model.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00003.safetensors",
571
- "model.vision_model.encoder.layers.9.attn.projection_layer.bias": "model-00001-of-00003.safetensors",
572
- "model.vision_model.encoder.layers.9.attn.projection_layer.weight": "model-00001-of-00003.safetensors",
573
- "model.vision_model.encoder.layers.9.attn.qkv.bias": "model-00001-of-00003.safetensors",
574
- "model.vision_model.encoder.layers.9.attn.qkv.weight": "model-00001-of-00003.safetensors",
 
 
 
 
575
  "model.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00003.safetensors",
576
  "model.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00003.safetensors",
577
  "model.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00003.safetensors",
@@ -580,19 +668,14 @@
580
  "model.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00003.safetensors",
581
  "model.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00003.safetensors",
582
  "model.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00003.safetensors",
583
- "model.vision_model.head.kv.bias": "model-00001-of-00003.safetensors",
584
- "model.vision_model.head.kv.weight": "model-00001-of-00003.safetensors",
585
- "model.vision_model.head.latent": "model-00001-of-00003.safetensors",
586
- "model.vision_model.head.layer_norm.bias": "model-00001-of-00003.safetensors",
587
- "model.vision_model.head.layer_norm.weight": "model-00001-of-00003.safetensors",
588
- "model.vision_model.head.mlp.fc1.bias": "model-00001-of-00003.safetensors",
589
- "model.vision_model.head.mlp.fc1.weight": "model-00001-of-00003.safetensors",
590
- "model.vision_model.head.mlp.fc2.bias": "model-00001-of-00003.safetensors",
591
- "model.vision_model.head.mlp.fc2.weight": "model-00001-of-00003.safetensors",
592
- "model.vision_model.head.projection_layer.bias": "model-00001-of-00003.safetensors",
593
- "model.vision_model.head.projection_layer.weight": "model-00001-of-00003.safetensors",
594
- "model.vision_model.head.q.bias": "model-00001-of-00003.safetensors",
595
- "model.vision_model.head.q.weight": "model-00001-of-00003.safetensors",
596
  "model.vision_model.post_layernorm.bias": "model-00001-of-00003.safetensors",
597
  "model.vision_model.post_layernorm.weight": "model-00001-of-00003.safetensors",
598
  "model.vqmodel.decoder.conv_in.bias": "model-00001-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 14815546646
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",
 
8
  "model.aligner.fc1.weight": "model-00001-of-00003.safetensors",
9
  "model.aligner.hidden_layers.0.bias": "model-00001-of-00003.safetensors",
10
  "model.aligner.hidden_layers.0.weight": "model-00001-of-00003.safetensors",
11
+ "model.generation_aligner.fc1.bias": "model-00001-of-00003.safetensors",
12
+ "model.generation_aligner.fc1.weight": "model-00001-of-00003.safetensors",
13
+ "model.generation_aligner.hidden_layers.0.bias": "model-00001-of-00003.safetensors",
14
+ "model.generation_aligner.hidden_layers.0.weight": "model-00001-of-00003.safetensors",
15
+ "model.generation_embeddings.weight": "model-00001-of-00003.safetensors",
16
+ "model.generation_head.proj_out.bias": "model-00001-of-00003.safetensors",
17
+ "model.generation_head.proj_out.weight": "model-00001-of-00003.safetensors",
18
+ "model.generation_head.vision_head.bias": "model-00001-of-00003.safetensors",
19
+ "model.generation_head.vision_head.weight": "model-00001-of-00003.safetensors",
20
  "model.language_model.embed_tokens.weight": "model-00001-of-00003.safetensors",
21
  "model.language_model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
22
  "model.language_model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
 
117
  "model.language_model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
118
  "model.language_model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
119
  "model.language_model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
120
+ "model.language_model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
121
+ "model.language_model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
122
  "model.language_model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
123
  "model.language_model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
124
+ "model.language_model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
125
  "model.language_model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
126
  "model.language_model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
127
  "model.language_model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
 
142
  "model.language_model.layers.20.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
143
  "model.language_model.layers.20.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
144
  "model.language_model.layers.20.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.language_model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
146
  "model.language_model.layers.20.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
147
  "model.language_model.layers.21.input_layernorm.weight": "model-00003-of-00003.safetensors",
148
  "model.language_model.layers.21.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
 
264
  "model.language_model.layers.7.input_layernorm.weight": "model-00002-of-00003.safetensors",
265
  "model.language_model.layers.7.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
266
  "model.language_model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.language_model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
268
  "model.language_model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
269
  "model.language_model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
270
  "model.language_model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
 
289
  "model.language_model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
290
  "model.language_model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
291
  "model.language_model.norm.weight": "model-00003-of-00003.safetensors",
292
+ "model.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00003.safetensors",
293
+ "model.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00003.safetensors",
294
+ "model.vision_model.embeddings.position_embedding.weight": "model-00001-of-00003.safetensors",
 
 
 
 
295
  "model.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00003.safetensors",
296
  "model.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00003.safetensors",
297
  "model.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
300
  "model.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00003.safetensors",
301
  "model.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00003.safetensors",
302
  "model.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00003.safetensors",
303
+ "model.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
304
+ "model.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
305
+ "model.vision_model.encoder.layers.0.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
306
+ "model.vision_model.encoder.layers.0.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
307
+ "model.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
308
+ "model.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
309
+ "model.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
310
+ "model.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
311
  "model.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00003.safetensors",
312
  "model.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00003.safetensors",
313
  "model.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
316
  "model.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00003.safetensors",
317
  "model.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00003.safetensors",
318
  "model.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00003.safetensors",
319
+ "model.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
320
+ "model.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
321
+ "model.vision_model.encoder.layers.1.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
322
+ "model.vision_model.encoder.layers.1.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
323
+ "model.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
324
+ "model.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
325
+ "model.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
326
+ "model.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
327
  "model.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00003.safetensors",
328
  "model.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00003.safetensors",
329
  "model.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
332
  "model.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00003.safetensors",
333
  "model.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00003.safetensors",
334
  "model.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00003.safetensors",
335
+ "model.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
336
+ "model.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
337
+ "model.vision_model.encoder.layers.10.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
338
+ "model.vision_model.encoder.layers.10.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
339
+ "model.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
340
+ "model.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
341
+ "model.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
342
+ "model.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
343
  "model.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00003.safetensors",
344
  "model.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00003.safetensors",
345
  "model.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
348
  "model.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00003.safetensors",
349
  "model.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00003.safetensors",
350
  "model.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00003.safetensors",
351
+ "model.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
352
+ "model.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
353
+ "model.vision_model.encoder.layers.11.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
354
+ "model.vision_model.encoder.layers.11.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
355
+ "model.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
356
+ "model.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
357
+ "model.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
358
+ "model.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
359
  "model.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00003.safetensors",
360
  "model.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00003.safetensors",
361
  "model.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
364
  "model.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00003.safetensors",
365
  "model.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00003.safetensors",
366
  "model.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00003.safetensors",
367
+ "model.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
368
+ "model.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
369
+ "model.vision_model.encoder.layers.12.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
370
+ "model.vision_model.encoder.layers.12.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
371
+ "model.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
372
+ "model.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
373
+ "model.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
374
+ "model.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
375
  "model.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00003.safetensors",
376
  "model.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00003.safetensors",
377
  "model.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
380
  "model.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00003.safetensors",
381
  "model.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00003.safetensors",
382
  "model.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00003.safetensors",
383
+ "model.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
384
+ "model.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
385
+ "model.vision_model.encoder.layers.13.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
386
+ "model.vision_model.encoder.layers.13.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
387
+ "model.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
388
+ "model.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
389
+ "model.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
390
+ "model.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
391
  "model.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00003.safetensors",
392
  "model.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00003.safetensors",
393
  "model.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
396
  "model.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00003.safetensors",
397
  "model.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00003.safetensors",
398
  "model.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00003.safetensors",
399
+ "model.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
400
+ "model.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
401
+ "model.vision_model.encoder.layers.14.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
402
+ "model.vision_model.encoder.layers.14.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
403
+ "model.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
404
+ "model.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
405
+ "model.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
406
+ "model.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
407
  "model.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00003.safetensors",
408
  "model.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00003.safetensors",
409
  "model.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
412
  "model.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00003.safetensors",
413
  "model.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00003.safetensors",
414
  "model.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00003.safetensors",
415
+ "model.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
416
+ "model.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
417
+ "model.vision_model.encoder.layers.15.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
418
+ "model.vision_model.encoder.layers.15.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
419
+ "model.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
420
+ "model.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
421
+ "model.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
422
+ "model.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
423
  "model.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00003.safetensors",
424
  "model.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00003.safetensors",
425
  "model.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
428
  "model.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00003.safetensors",
429
  "model.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00003.safetensors",
430
  "model.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00003.safetensors",
431
+ "model.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
432
+ "model.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
433
+ "model.vision_model.encoder.layers.16.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
434
+ "model.vision_model.encoder.layers.16.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
435
+ "model.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
436
+ "model.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
437
+ "model.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
438
+ "model.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
439
  "model.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00003.safetensors",
440
  "model.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00003.safetensors",
441
  "model.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
444
  "model.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00003.safetensors",
445
  "model.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00003.safetensors",
446
  "model.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00003.safetensors",
447
+ "model.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
448
+ "model.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
449
+ "model.vision_model.encoder.layers.17.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
450
+ "model.vision_model.encoder.layers.17.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
451
+ "model.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
452
+ "model.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
453
+ "model.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
454
+ "model.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
455
  "model.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00003.safetensors",
456
  "model.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00003.safetensors",
457
  "model.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
460
  "model.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00003.safetensors",
461
  "model.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00003.safetensors",
462
  "model.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00003.safetensors",
463
+ "model.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
464
+ "model.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
465
+ "model.vision_model.encoder.layers.18.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
466
+ "model.vision_model.encoder.layers.18.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
467
+ "model.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
468
+ "model.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
469
+ "model.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
470
+ "model.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
471
  "model.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00003.safetensors",
472
  "model.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00003.safetensors",
473
  "model.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
476
  "model.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00003.safetensors",
477
  "model.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00003.safetensors",
478
  "model.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00003.safetensors",
479
+ "model.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
480
+ "model.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
481
+ "model.vision_model.encoder.layers.19.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
482
+ "model.vision_model.encoder.layers.19.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
483
+ "model.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
484
+ "model.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
485
+ "model.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
486
+ "model.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
487
  "model.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00003.safetensors",
488
  "model.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00003.safetensors",
489
  "model.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
492
  "model.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00003.safetensors",
493
  "model.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00003.safetensors",
494
  "model.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00003.safetensors",
495
+ "model.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
496
+ "model.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
497
+ "model.vision_model.encoder.layers.2.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
498
+ "model.vision_model.encoder.layers.2.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
499
+ "model.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
500
+ "model.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
501
+ "model.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
502
+ "model.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
503
  "model.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00003.safetensors",
504
  "model.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00003.safetensors",
505
  "model.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
508
  "model.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00003.safetensors",
509
  "model.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00003.safetensors",
510
  "model.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00003.safetensors",
511
+ "model.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
512
+ "model.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
513
+ "model.vision_model.encoder.layers.20.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
514
+ "model.vision_model.encoder.layers.20.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
515
+ "model.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
516
+ "model.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
517
+ "model.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
518
+ "model.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
519
  "model.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00003.safetensors",
520
  "model.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00003.safetensors",
521
  "model.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
524
  "model.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00003.safetensors",
525
  "model.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00003.safetensors",
526
  "model.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00003.safetensors",
527
+ "model.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
528
+ "model.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
529
+ "model.vision_model.encoder.layers.21.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
530
+ "model.vision_model.encoder.layers.21.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
531
+ "model.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
532
+ "model.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
533
+ "model.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
534
+ "model.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
535
  "model.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00003.safetensors",
536
  "model.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00003.safetensors",
537
  "model.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
540
  "model.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00003.safetensors",
541
  "model.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00003.safetensors",
542
  "model.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00003.safetensors",
543
+ "model.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
544
+ "model.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
545
+ "model.vision_model.encoder.layers.22.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
546
+ "model.vision_model.encoder.layers.22.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
547
+ "model.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
548
+ "model.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
549
+ "model.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
550
+ "model.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
551
  "model.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00003.safetensors",
552
  "model.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00003.safetensors",
553
  "model.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
556
  "model.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00003.safetensors",
557
  "model.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00003.safetensors",
558
  "model.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00003.safetensors",
559
+ "model.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
560
+ "model.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
561
+ "model.vision_model.encoder.layers.23.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
562
+ "model.vision_model.encoder.layers.23.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
563
+ "model.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
564
+ "model.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
565
+ "model.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
566
+ "model.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
567
  "model.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00003.safetensors",
568
  "model.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00003.safetensors",
569
  "model.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
572
  "model.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00003.safetensors",
573
  "model.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00003.safetensors",
574
  "model.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00003.safetensors",
575
+ "model.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
576
+ "model.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
577
+ "model.vision_model.encoder.layers.3.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
578
+ "model.vision_model.encoder.layers.3.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
579
+ "model.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
580
+ "model.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
581
+ "model.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
582
+ "model.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
583
  "model.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00003.safetensors",
584
  "model.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00003.safetensors",
585
  "model.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
588
  "model.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00003.safetensors",
589
  "model.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00003.safetensors",
590
  "model.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00003.safetensors",
591
+ "model.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
592
+ "model.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
593
+ "model.vision_model.encoder.layers.4.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
594
+ "model.vision_model.encoder.layers.4.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
595
+ "model.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
596
+ "model.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
597
+ "model.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
598
+ "model.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
599
  "model.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00003.safetensors",
600
  "model.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00003.safetensors",
601
  "model.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
604
  "model.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00003.safetensors",
605
  "model.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00003.safetensors",
606
  "model.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00003.safetensors",
607
+ "model.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
608
+ "model.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
609
+ "model.vision_model.encoder.layers.5.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
610
+ "model.vision_model.encoder.layers.5.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
611
+ "model.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
612
+ "model.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
613
+ "model.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
614
+ "model.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
615
  "model.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00003.safetensors",
616
  "model.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00003.safetensors",
617
  "model.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
620
  "model.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00003.safetensors",
621
  "model.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00003.safetensors",
622
  "model.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00003.safetensors",
623
+ "model.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
624
+ "model.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
625
+ "model.vision_model.encoder.layers.6.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
626
+ "model.vision_model.encoder.layers.6.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
627
+ "model.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
628
+ "model.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
629
+ "model.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
630
+ "model.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
631
  "model.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00003.safetensors",
632
  "model.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00003.safetensors",
633
  "model.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
636
  "model.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00003.safetensors",
637
  "model.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00003.safetensors",
638
  "model.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00003.safetensors",
639
+ "model.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
640
+ "model.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
641
+ "model.vision_model.encoder.layers.7.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
642
+ "model.vision_model.encoder.layers.7.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
643
+ "model.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
644
+ "model.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
645
+ "model.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
646
+ "model.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
647
  "model.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00003.safetensors",
648
  "model.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00003.safetensors",
649
  "model.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
652
  "model.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00003.safetensors",
653
  "model.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00003.safetensors",
654
  "model.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00003.safetensors",
655
+ "model.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
656
+ "model.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
657
+ "model.vision_model.encoder.layers.8.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
658
+ "model.vision_model.encoder.layers.8.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
659
+ "model.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
660
+ "model.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
661
+ "model.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
662
+ "model.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
663
  "model.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00003.safetensors",
664
  "model.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00003.safetensors",
665
  "model.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00003.safetensors",
 
668
  "model.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00003.safetensors",
669
  "model.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00003.safetensors",
670
  "model.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00003.safetensors",
671
+ "model.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
672
+ "model.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
673
+ "model.vision_model.encoder.layers.9.self_attn.projection_layer.bias": "model-00001-of-00003.safetensors",
674
+ "model.vision_model.encoder.layers.9.self_attn.projection_layer.weight": "model-00001-of-00003.safetensors",
675
+ "model.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
676
+ "model.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
677
+ "model.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00003.safetensors",
678
+ "model.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
 
 
 
 
 
679
  "model.vision_model.post_layernorm.bias": "model-00001-of-00003.safetensors",
680
  "model.vision_model.post_layernorm.weight": "model-00001-of-00003.safetensors",
681
  "model.vqmodel.decoder.conv_in.bias": "model-00001-of-00003.safetensors",