kcz358 commited on
Commit
ce82502
1 Parent(s): e7fdc94

Upload config

Browse files
Files changed (1) hide show
  1. config.json +628 -0
config.json ADDED
@@ -0,0 +1,628 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "llavanext-google_siglip-so400m-patch14-384-Qwen_Qwen2-7B-Instruct-blip558k_pretrain_mid_stage_ft",
3
+ "architectures": [
4
+ "LlavaQwenForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "ignore_index": -100,
12
+ "image_aspect_ratio": "anyres_max_9",
13
+ "image_crop_resolution": null,
14
+ "image_grid_pinpoints": [
15
+ [
16
+ 384,
17
+ 384
18
+ ],
19
+ [
20
+ 384,
21
+ 768
22
+ ],
23
+ [
24
+ 384,
25
+ 1152
26
+ ],
27
+ [
28
+ 384,
29
+ 1536
30
+ ],
31
+ [
32
+ 768,
33
+ 768
34
+ ],
35
+ [
36
+ 384,
37
+ 1920
38
+ ],
39
+ [
40
+ 768,
41
+ 768
42
+ ],
43
+ [
44
+ 384,
45
+ 2304
46
+ ],
47
+ [
48
+ 768,
49
+ 1152
50
+ ],
51
+ [
52
+ 384,
53
+ 2688
54
+ ],
55
+ [
56
+ 768,
57
+ 1152
58
+ ],
59
+ [
60
+ 384,
61
+ 3072
62
+ ],
63
+ [
64
+ 768,
65
+ 1536
66
+ ],
67
+ [
68
+ 384,
69
+ 3456
70
+ ],
71
+ [
72
+ 768,
73
+ 1536
74
+ ],
75
+ [
76
+ 1152,
77
+ 1152
78
+ ],
79
+ [
80
+ 768,
81
+ 1920
82
+ ],
83
+ [
84
+ 1152,
85
+ 1152
86
+ ],
87
+ [
88
+ 768,
89
+ 1920
90
+ ],
91
+ [
92
+ 1152,
93
+ 1152
94
+ ],
95
+ [
96
+ 768,
97
+ 2304
98
+ ],
99
+ [
100
+ 1152,
101
+ 1536
102
+ ],
103
+ [
104
+ 768,
105
+ 2304
106
+ ],
107
+ [
108
+ 1152,
109
+ 1536
110
+ ],
111
+ [
112
+ 768,
113
+ 2688
114
+ ],
115
+ [
116
+ 1152,
117
+ 1536
118
+ ],
119
+ [
120
+ 768,
121
+ 2688
122
+ ],
123
+ [
124
+ 1152,
125
+ 1920
126
+ ],
127
+ [
128
+ 768,
129
+ 3072
130
+ ],
131
+ [
132
+ 1152,
133
+ 1920
134
+ ],
135
+ [
136
+ 1536,
137
+ 1536
138
+ ],
139
+ [
140
+ 768,
141
+ 3072
142
+ ],
143
+ [
144
+ 1152,
145
+ 1920
146
+ ],
147
+ [
148
+ 1536,
149
+ 1536
150
+ ],
151
+ [
152
+ 768,
153
+ 3456
154
+ ],
155
+ [
156
+ 1152,
157
+ 2304
158
+ ],
159
+ [
160
+ 1536,
161
+ 1536
162
+ ],
163
+ [
164
+ 768,
165
+ 3456
166
+ ],
167
+ [
168
+ 1152,
169
+ 2304
170
+ ],
171
+ [
172
+ 1536,
173
+ 1536
174
+ ],
175
+ [
176
+ 768,
177
+ 3840
178
+ ],
179
+ [
180
+ 1152,
181
+ 2304
182
+ ],
183
+ [
184
+ 1536,
185
+ 1920
186
+ ],
187
+ [
188
+ 768,
189
+ 3840
190
+ ],
191
+ [
192
+ 1152,
193
+ 2688
194
+ ],
195
+ [
196
+ 1536,
197
+ 1920
198
+ ],
199
+ [
200
+ 768,
201
+ 4224
202
+ ],
203
+ [
204
+ 1152,
205
+ 2688
206
+ ],
207
+ [
208
+ 1536,
209
+ 1920
210
+ ],
211
+ [
212
+ 768,
213
+ 4224
214
+ ],
215
+ [
216
+ 1152,
217
+ 2688
218
+ ],
219
+ [
220
+ 1536,
221
+ 1920
222
+ ],
223
+ [
224
+ 768,
225
+ 4608
226
+ ],
227
+ [
228
+ 1152,
229
+ 3072
230
+ ],
231
+ [
232
+ 1536,
233
+ 2304
234
+ ],
235
+ [
236
+ 768,
237
+ 4608
238
+ ],
239
+ [
240
+ 1152,
241
+ 3072
242
+ ],
243
+ [
244
+ 1536,
245
+ 2304
246
+ ],
247
+ [
248
+ 1920,
249
+ 1920
250
+ ],
251
+ [
252
+ 768,
253
+ 4992
254
+ ],
255
+ [
256
+ 1152,
257
+ 3072
258
+ ],
259
+ [
260
+ 1536,
261
+ 2304
262
+ ],
263
+ [
264
+ 1920,
265
+ 1920
266
+ ],
267
+ [
268
+ 768,
269
+ 4992
270
+ ],
271
+ [
272
+ 1152,
273
+ 3456
274
+ ],
275
+ [
276
+ 1536,
277
+ 2304
278
+ ],
279
+ [
280
+ 1920,
281
+ 1920
282
+ ],
283
+ [
284
+ 768,
285
+ 5376
286
+ ],
287
+ [
288
+ 1152,
289
+ 3456
290
+ ],
291
+ [
292
+ 1536,
293
+ 2688
294
+ ],
295
+ [
296
+ 1920,
297
+ 1920
298
+ ],
299
+ [
300
+ 768,
301
+ 5376
302
+ ],
303
+ [
304
+ 1152,
305
+ 3456
306
+ ],
307
+ [
308
+ 1536,
309
+ 2688
310
+ ],
311
+ [
312
+ 1920,
313
+ 1920
314
+ ],
315
+ [
316
+ 768,
317
+ 5760
318
+ ],
319
+ [
320
+ 1152,
321
+ 3840
322
+ ],
323
+ [
324
+ 1536,
325
+ 2688
326
+ ],
327
+ [
328
+ 1920,
329
+ 2304
330
+ ],
331
+ [
332
+ 768,
333
+ 5760
334
+ ],
335
+ [
336
+ 1152,
337
+ 3840
338
+ ],
339
+ [
340
+ 1536,
341
+ 2688
342
+ ],
343
+ [
344
+ 1920,
345
+ 2304
346
+ ],
347
+ [
348
+ 768,
349
+ 6144
350
+ ],
351
+ [
352
+ 1152,
353
+ 3840
354
+ ],
355
+ [
356
+ 1536,
357
+ 3072
358
+ ],
359
+ [
360
+ 1920,
361
+ 2304
362
+ ],
363
+ [
364
+ 768,
365
+ 6144
366
+ ],
367
+ [
368
+ 1152,
369
+ 4224
370
+ ],
371
+ [
372
+ 1536,
373
+ 3072
374
+ ],
375
+ [
376
+ 1920,
377
+ 2304
378
+ ],
379
+ [
380
+ 768,
381
+ 6528
382
+ ],
383
+ [
384
+ 1152,
385
+ 4224
386
+ ],
387
+ [
388
+ 1536,
389
+ 3072
390
+ ],
391
+ [
392
+ 1920,
393
+ 2304
394
+ ],
395
+ [
396
+ 768,
397
+ 6528
398
+ ],
399
+ [
400
+ 1152,
401
+ 4224
402
+ ],
403
+ [
404
+ 1536,
405
+ 3072
406
+ ],
407
+ [
408
+ 1920,
409
+ 2688
410
+ ],
411
+ [
412
+ 768,
413
+ 6912
414
+ ],
415
+ [
416
+ 1152,
417
+ 4608
418
+ ],
419
+ [
420
+ 1536,
421
+ 3456
422
+ ],
423
+ [
424
+ 1920,
425
+ 2688
426
+ ],
427
+ [
428
+ 2304,
429
+ 2304
430
+ ]
431
+ ],
432
+ "image_split_resolution": null,
433
+ "image_token_index": 151646,
434
+ "initializer_range": 0.02,
435
+ "intermediate_size": 18944,
436
+ "max_position_embeddings": 32768,
437
+ "max_window_layers": 28,
438
+ "mm_hidden_size": 1152,
439
+ "mm_patch_merge_type": "spatial_unpad",
440
+ "mm_projector_lr": null,
441
+ "mm_projector_type": "mlp2x_gelu",
442
+ "mm_resampler_type": null,
443
+ "mm_tunable_parts": "mm_vision_tower,mm_mlp_adapter,mm_language_model",
444
+ "mm_use_im_patch_token": false,
445
+ "mm_use_im_start_end": false,
446
+ "mm_vision_select_feature": "patch",
447
+ "mm_vision_select_layer": -2,
448
+ "mm_vision_tower": "google/siglip-so400m-patch14-384",
449
+ "mm_vision_tower_lr": 2e-06,
450
+ "model_type": "llava",
451
+ "num_attention_heads": 28,
452
+ "num_hidden_layers": 28,
453
+ "num_key_value_heads": 4,
454
+ "pos_skipping_range": 4096,
455
+ "projector_hidden_act": "gelu",
456
+ "rms_norm_eps": 1e-06,
457
+ "rope_scaling": null,
458
+ "rope_theta": 1000000.0,
459
+ "sliding_window": 131072,
460
+ "text_config": {
461
+ "model_type": "llama"
462
+ },
463
+ "tie_word_embeddings": false,
464
+ "tokenizer_model_max_length": 32768,
465
+ "tokenizer_padding_side": "right",
466
+ "torch_dtype": "bfloat16",
467
+ "transformers_version": "4.42.3",
468
+ "use_cache": true,
469
+ "use_mm_proj": true,
470
+ "use_pos_skipping": false,
471
+ "use_sliding_window": false,
472
+ "vision_config": {
473
+ "_name_or_path": "google/siglip-so400m-patch14-384",
474
+ "architectures": [
475
+ "SiglipModel"
476
+ ],
477
+ "model_type": "siglip",
478
+ "text_config": {
479
+ "_name_or_path": "",
480
+ "add_cross_attention": false,
481
+ "architectures": null,
482
+ "attention_dropout": 0.0,
483
+ "bad_words_ids": null,
484
+ "begin_suppress_tokens": null,
485
+ "bos_token_id": 49406,
486
+ "chunk_size_feed_forward": 0,
487
+ "cross_attention_hidden_size": null,
488
+ "decoder_start_token_id": null,
489
+ "diversity_penalty": 0.0,
490
+ "do_sample": false,
491
+ "early_stopping": false,
492
+ "encoder_no_repeat_ngram_size": 0,
493
+ "eos_token_id": 49407,
494
+ "exponential_decay_length_penalty": null,
495
+ "finetuning_task": null,
496
+ "forced_bos_token_id": null,
497
+ "forced_eos_token_id": null,
498
+ "hidden_act": "gelu_pytorch_tanh",
499
+ "hidden_size": 1152,
500
+ "id2label": {
501
+ "0": "LABEL_0",
502
+ "1": "LABEL_1"
503
+ },
504
+ "intermediate_size": 4304,
505
+ "is_decoder": false,
506
+ "is_encoder_decoder": false,
507
+ "label2id": {
508
+ "LABEL_0": 0,
509
+ "LABEL_1": 1
510
+ },
511
+ "layer_norm_eps": 1e-06,
512
+ "length_penalty": 1.0,
513
+ "max_length": 20,
514
+ "max_position_embeddings": 64,
515
+ "min_length": 0,
516
+ "model_type": "siglip_text_model",
517
+ "no_repeat_ngram_size": 0,
518
+ "num_attention_heads": 16,
519
+ "num_beam_groups": 1,
520
+ "num_beams": 1,
521
+ "num_hidden_layers": 27,
522
+ "num_return_sequences": 1,
523
+ "output_attentions": false,
524
+ "output_hidden_states": false,
525
+ "output_scores": false,
526
+ "pad_token_id": 1,
527
+ "prefix": null,
528
+ "problem_type": null,
529
+ "pruned_heads": {},
530
+ "remove_invalid_values": false,
531
+ "repetition_penalty": 1.0,
532
+ "return_dict": true,
533
+ "return_dict_in_generate": false,
534
+ "sep_token_id": null,
535
+ "suppress_tokens": null,
536
+ "task_specific_params": null,
537
+ "temperature": 1.0,
538
+ "tf_legacy_loss": false,
539
+ "tie_encoder_decoder": false,
540
+ "tie_word_embeddings": true,
541
+ "tokenizer_class": null,
542
+ "top_k": 50,
543
+ "top_p": 1.0,
544
+ "torch_dtype": null,
545
+ "torchscript": false,
546
+ "typical_p": 1.0,
547
+ "use_bfloat16": false,
548
+ "vocab_size": 32000
549
+ },
550
+ "torch_dtype": "float32",
551
+ "vision_config": {
552
+ "_name_or_path": "",
553
+ "add_cross_attention": false,
554
+ "architectures": null,
555
+ "attention_dropout": 0.0,
556
+ "bad_words_ids": null,
557
+ "begin_suppress_tokens": null,
558
+ "bos_token_id": null,
559
+ "chunk_size_feed_forward": 0,
560
+ "cross_attention_hidden_size": null,
561
+ "decoder_start_token_id": null,
562
+ "diversity_penalty": 0.0,
563
+ "do_sample": false,
564
+ "early_stopping": false,
565
+ "encoder_no_repeat_ngram_size": 0,
566
+ "eos_token_id": null,
567
+ "exponential_decay_length_penalty": null,
568
+ "finetuning_task": null,
569
+ "forced_bos_token_id": null,
570
+ "forced_eos_token_id": null,
571
+ "hidden_act": "gelu_pytorch_tanh",
572
+ "hidden_size": 1152,
573
+ "id2label": {
574
+ "0": "LABEL_0",
575
+ "1": "LABEL_1"
576
+ },
577
+ "image_size": 384,
578
+ "intermediate_size": 4304,
579
+ "is_decoder": false,
580
+ "is_encoder_decoder": false,
581
+ "label2id": {
582
+ "LABEL_0": 0,
583
+ "LABEL_1": 1
584
+ },
585
+ "layer_norm_eps": 1e-06,
586
+ "length_penalty": 1.0,
587
+ "max_length": 20,
588
+ "min_length": 0,
589
+ "model_type": "siglip_vision_model",
590
+ "no_repeat_ngram_size": 0,
591
+ "num_attention_heads": 16,
592
+ "num_beam_groups": 1,
593
+ "num_beams": 1,
594
+ "num_channels": 3,
595
+ "num_hidden_layers": 27,
596
+ "num_return_sequences": 1,
597
+ "output_attentions": false,
598
+ "output_hidden_states": false,
599
+ "output_scores": false,
600
+ "pad_token_id": null,
601
+ "patch_size": 14,
602
+ "prefix": null,
603
+ "problem_type": null,
604
+ "pruned_heads": {},
605
+ "remove_invalid_values": false,
606
+ "repetition_penalty": 1.0,
607
+ "return_dict": true,
608
+ "return_dict_in_generate": false,
609
+ "sep_token_id": null,
610
+ "suppress_tokens": null,
611
+ "task_specific_params": null,
612
+ "temperature": 1.0,
613
+ "tf_legacy_loss": false,
614
+ "tie_encoder_decoder": false,
615
+ "tie_word_embeddings": true,
616
+ "tokenizer_class": null,
617
+ "top_k": 50,
618
+ "top_p": 1.0,
619
+ "torch_dtype": null,
620
+ "torchscript": false,
621
+ "typical_p": 1.0,
622
+ "use_bfloat16": false
623
+ }
624
+ },
625
+ "vision_feature_layer": -2,
626
+ "vision_feature_select_strategy": "default",
627
+ "vision_tower_pretrained": null
628
+ }