MoonstoneF
commited on
Commit
•
8df4359
1
Parent(s):
6fb10ce
Trained on 6291 without hyperparameter tuning
Browse files- README.md +69 -0
- added_tokens.json +1037 -0
- config.json +163 -0
- generation_config.json +7 -0
- model-00001-of-00002.safetensors +3 -0
- model-00002-of-00002.safetensors +3 -0
- model.safetensors.index.json +892 -0
- preprocessor_config.json +29 -0
- processor_config.json +4 -0
- sentencepiece.bpe.model +3 -0
- special_tokens_map.json +58 -0
- tokenizer.json +0 -0
- tokenizer_config.json +0 -0
- training_args.bin +3 -0
README.md
ADDED
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
base_model: microsoft/kosmos-2-patch14-224
|
4 |
+
tags:
|
5 |
+
- generated_from_trainer
|
6 |
+
model-index:
|
7 |
+
- name: kosm-checkpoint
|
8 |
+
results: []
|
9 |
+
---
|
10 |
+
|
11 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
12 |
+
should probably proofread and complete it, then remove this comment. -->
|
13 |
+
|
14 |
+
# kosm-checkpoint
|
15 |
+
|
16 |
+
This model is a fine-tuned version of [microsoft/kosmos-2-patch14-224](https://huggingface.co/microsoft/kosmos-2-patch14-224) on the None dataset.
|
17 |
+
It achieves the following results on the evaluation set:
|
18 |
+
- Loss: 0.0715
|
19 |
+
|
20 |
+
## Model description
|
21 |
+
|
22 |
+
More information needed
|
23 |
+
|
24 |
+
## Intended uses & limitations
|
25 |
+
|
26 |
+
More information needed
|
27 |
+
|
28 |
+
## Training and evaluation data
|
29 |
+
|
30 |
+
More information needed
|
31 |
+
|
32 |
+
## Training procedure
|
33 |
+
|
34 |
+
### Training hyperparameters
|
35 |
+
|
36 |
+
The following hyperparameters were used during training:
|
37 |
+
- learning_rate: 1e-05
|
38 |
+
- train_batch_size: 2
|
39 |
+
- eval_batch_size: 2
|
40 |
+
- seed: 42
|
41 |
+
- gradient_accumulation_steps: 2
|
42 |
+
- total_train_batch_size: 4
|
43 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
44 |
+
- lr_scheduler_type: linear
|
45 |
+
- lr_scheduler_warmup_ratio: 0.1
|
46 |
+
- training_steps: 1000
|
47 |
+
|
48 |
+
### Training results
|
49 |
+
|
50 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
51 |
+
|:-------------:|:------:|:----:|:---------------:|
|
52 |
+
| 0.0711 | 0.0908 | 100 | 0.1326 |
|
53 |
+
| 0.1093 | 0.1817 | 200 | 0.0949 |
|
54 |
+
| 0.0946 | 0.2725 | 300 | 0.0908 |
|
55 |
+
| 0.0907 | 0.3633 | 400 | 0.0848 |
|
56 |
+
| 0.0829 | 0.4541 | 500 | 0.0812 |
|
57 |
+
| 0.082 | 0.5450 | 600 | 0.0771 |
|
58 |
+
| 0.0792 | 0.6358 | 700 | 0.0759 |
|
59 |
+
| 0.0773 | 0.7266 | 800 | 0.0737 |
|
60 |
+
| 0.0736 | 0.8174 | 900 | 0.0723 |
|
61 |
+
| 0.0753 | 0.9083 | 1000 | 0.0715 |
|
62 |
+
|
63 |
+
|
64 |
+
### Framework versions
|
65 |
+
|
66 |
+
- Transformers 4.42.4
|
67 |
+
- Pytorch 2.1.2+cu121
|
68 |
+
- Datasets 2.15.0
|
69 |
+
- Tokenizers 0.19.1
|
added_tokens.json
ADDED
@@ -0,0 +1,1037 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</chunk>": 64005,
|
3 |
+
"</delimiter_of_multi_objects/>": 64011,
|
4 |
+
"</doc>": 64002,
|
5 |
+
"</image>": 64004,
|
6 |
+
"</line>": 64006,
|
7 |
+
"</object>": 64010,
|
8 |
+
"</phrase>": 64008,
|
9 |
+
"<grounding>": 64012,
|
10 |
+
"<image>": 64003,
|
11 |
+
"<object>": 64009,
|
12 |
+
"<patch_index_0000>": 64013,
|
13 |
+
"<patch_index_0001>": 64014,
|
14 |
+
"<patch_index_0002>": 64015,
|
15 |
+
"<patch_index_0003>": 64016,
|
16 |
+
"<patch_index_0004>": 64017,
|
17 |
+
"<patch_index_0005>": 64018,
|
18 |
+
"<patch_index_0006>": 64019,
|
19 |
+
"<patch_index_0007>": 64020,
|
20 |
+
"<patch_index_0008>": 64021,
|
21 |
+
"<patch_index_0009>": 64022,
|
22 |
+
"<patch_index_0010>": 64023,
|
23 |
+
"<patch_index_0011>": 64024,
|
24 |
+
"<patch_index_0012>": 64025,
|
25 |
+
"<patch_index_0013>": 64026,
|
26 |
+
"<patch_index_0014>": 64027,
|
27 |
+
"<patch_index_0015>": 64028,
|
28 |
+
"<patch_index_0016>": 64029,
|
29 |
+
"<patch_index_0017>": 64030,
|
30 |
+
"<patch_index_0018>": 64031,
|
31 |
+
"<patch_index_0019>": 64032,
|
32 |
+
"<patch_index_0020>": 64033,
|
33 |
+
"<patch_index_0021>": 64034,
|
34 |
+
"<patch_index_0022>": 64035,
|
35 |
+
"<patch_index_0023>": 64036,
|
36 |
+
"<patch_index_0024>": 64037,
|
37 |
+
"<patch_index_0025>": 64038,
|
38 |
+
"<patch_index_0026>": 64039,
|
39 |
+
"<patch_index_0027>": 64040,
|
40 |
+
"<patch_index_0028>": 64041,
|
41 |
+
"<patch_index_0029>": 64042,
|
42 |
+
"<patch_index_0030>": 64043,
|
43 |
+
"<patch_index_0031>": 64044,
|
44 |
+
"<patch_index_0032>": 64045,
|
45 |
+
"<patch_index_0033>": 64046,
|
46 |
+
"<patch_index_0034>": 64047,
|
47 |
+
"<patch_index_0035>": 64048,
|
48 |
+
"<patch_index_0036>": 64049,
|
49 |
+
"<patch_index_0037>": 64050,
|
50 |
+
"<patch_index_0038>": 64051,
|
51 |
+
"<patch_index_0039>": 64052,
|
52 |
+
"<patch_index_0040>": 64053,
|
53 |
+
"<patch_index_0041>": 64054,
|
54 |
+
"<patch_index_0042>": 64055,
|
55 |
+
"<patch_index_0043>": 64056,
|
56 |
+
"<patch_index_0044>": 64057,
|
57 |
+
"<patch_index_0045>": 64058,
|
58 |
+
"<patch_index_0046>": 64059,
|
59 |
+
"<patch_index_0047>": 64060,
|
60 |
+
"<patch_index_0048>": 64061,
|
61 |
+
"<patch_index_0049>": 64062,
|
62 |
+
"<patch_index_0050>": 64063,
|
63 |
+
"<patch_index_0051>": 64064,
|
64 |
+
"<patch_index_0052>": 64065,
|
65 |
+
"<patch_index_0053>": 64066,
|
66 |
+
"<patch_index_0054>": 64067,
|
67 |
+
"<patch_index_0055>": 64068,
|
68 |
+
"<patch_index_0056>": 64069,
|
69 |
+
"<patch_index_0057>": 64070,
|
70 |
+
"<patch_index_0058>": 64071,
|
71 |
+
"<patch_index_0059>": 64072,
|
72 |
+
"<patch_index_0060>": 64073,
|
73 |
+
"<patch_index_0061>": 64074,
|
74 |
+
"<patch_index_0062>": 64075,
|
75 |
+
"<patch_index_0063>": 64076,
|
76 |
+
"<patch_index_0064>": 64077,
|
77 |
+
"<patch_index_0065>": 64078,
|
78 |
+
"<patch_index_0066>": 64079,
|
79 |
+
"<patch_index_0067>": 64080,
|
80 |
+
"<patch_index_0068>": 64081,
|
81 |
+
"<patch_index_0069>": 64082,
|
82 |
+
"<patch_index_0070>": 64083,
|
83 |
+
"<patch_index_0071>": 64084,
|
84 |
+
"<patch_index_0072>": 64085,
|
85 |
+
"<patch_index_0073>": 64086,
|
86 |
+
"<patch_index_0074>": 64087,
|
87 |
+
"<patch_index_0075>": 64088,
|
88 |
+
"<patch_index_0076>": 64089,
|
89 |
+
"<patch_index_0077>": 64090,
|
90 |
+
"<patch_index_0078>": 64091,
|
91 |
+
"<patch_index_0079>": 64092,
|
92 |
+
"<patch_index_0080>": 64093,
|
93 |
+
"<patch_index_0081>": 64094,
|
94 |
+
"<patch_index_0082>": 64095,
|
95 |
+
"<patch_index_0083>": 64096,
|
96 |
+
"<patch_index_0084>": 64097,
|
97 |
+
"<patch_index_0085>": 64098,
|
98 |
+
"<patch_index_0086>": 64099,
|
99 |
+
"<patch_index_0087>": 64100,
|
100 |
+
"<patch_index_0088>": 64101,
|
101 |
+
"<patch_index_0089>": 64102,
|
102 |
+
"<patch_index_0090>": 64103,
|
103 |
+
"<patch_index_0091>": 64104,
|
104 |
+
"<patch_index_0092>": 64105,
|
105 |
+
"<patch_index_0093>": 64106,
|
106 |
+
"<patch_index_0094>": 64107,
|
107 |
+
"<patch_index_0095>": 64108,
|
108 |
+
"<patch_index_0096>": 64109,
|
109 |
+
"<patch_index_0097>": 64110,
|
110 |
+
"<patch_index_0098>": 64111,
|
111 |
+
"<patch_index_0099>": 64112,
|
112 |
+
"<patch_index_0100>": 64113,
|
113 |
+
"<patch_index_0101>": 64114,
|
114 |
+
"<patch_index_0102>": 64115,
|
115 |
+
"<patch_index_0103>": 64116,
|
116 |
+
"<patch_index_0104>": 64117,
|
117 |
+
"<patch_index_0105>": 64118,
|
118 |
+
"<patch_index_0106>": 64119,
|
119 |
+
"<patch_index_0107>": 64120,
|
120 |
+
"<patch_index_0108>": 64121,
|
121 |
+
"<patch_index_0109>": 64122,
|
122 |
+
"<patch_index_0110>": 64123,
|
123 |
+
"<patch_index_0111>": 64124,
|
124 |
+
"<patch_index_0112>": 64125,
|
125 |
+
"<patch_index_0113>": 64126,
|
126 |
+
"<patch_index_0114>": 64127,
|
127 |
+
"<patch_index_0115>": 64128,
|
128 |
+
"<patch_index_0116>": 64129,
|
129 |
+
"<patch_index_0117>": 64130,
|
130 |
+
"<patch_index_0118>": 64131,
|
131 |
+
"<patch_index_0119>": 64132,
|
132 |
+
"<patch_index_0120>": 64133,
|
133 |
+
"<patch_index_0121>": 64134,
|
134 |
+
"<patch_index_0122>": 64135,
|
135 |
+
"<patch_index_0123>": 64136,
|
136 |
+
"<patch_index_0124>": 64137,
|
137 |
+
"<patch_index_0125>": 64138,
|
138 |
+
"<patch_index_0126>": 64139,
|
139 |
+
"<patch_index_0127>": 64140,
|
140 |
+
"<patch_index_0128>": 64141,
|
141 |
+
"<patch_index_0129>": 64142,
|
142 |
+
"<patch_index_0130>": 64143,
|
143 |
+
"<patch_index_0131>": 64144,
|
144 |
+
"<patch_index_0132>": 64145,
|
145 |
+
"<patch_index_0133>": 64146,
|
146 |
+
"<patch_index_0134>": 64147,
|
147 |
+
"<patch_index_0135>": 64148,
|
148 |
+
"<patch_index_0136>": 64149,
|
149 |
+
"<patch_index_0137>": 64150,
|
150 |
+
"<patch_index_0138>": 64151,
|
151 |
+
"<patch_index_0139>": 64152,
|
152 |
+
"<patch_index_0140>": 64153,
|
153 |
+
"<patch_index_0141>": 64154,
|
154 |
+
"<patch_index_0142>": 64155,
|
155 |
+
"<patch_index_0143>": 64156,
|
156 |
+
"<patch_index_0144>": 64157,
|
157 |
+
"<patch_index_0145>": 64158,
|
158 |
+
"<patch_index_0146>": 64159,
|
159 |
+
"<patch_index_0147>": 64160,
|
160 |
+
"<patch_index_0148>": 64161,
|
161 |
+
"<patch_index_0149>": 64162,
|
162 |
+
"<patch_index_0150>": 64163,
|
163 |
+
"<patch_index_0151>": 64164,
|
164 |
+
"<patch_index_0152>": 64165,
|
165 |
+
"<patch_index_0153>": 64166,
|
166 |
+
"<patch_index_0154>": 64167,
|
167 |
+
"<patch_index_0155>": 64168,
|
168 |
+
"<patch_index_0156>": 64169,
|
169 |
+
"<patch_index_0157>": 64170,
|
170 |
+
"<patch_index_0158>": 64171,
|
171 |
+
"<patch_index_0159>": 64172,
|
172 |
+
"<patch_index_0160>": 64173,
|
173 |
+
"<patch_index_0161>": 64174,
|
174 |
+
"<patch_index_0162>": 64175,
|
175 |
+
"<patch_index_0163>": 64176,
|
176 |
+
"<patch_index_0164>": 64177,
|
177 |
+
"<patch_index_0165>": 64178,
|
178 |
+
"<patch_index_0166>": 64179,
|
179 |
+
"<patch_index_0167>": 64180,
|
180 |
+
"<patch_index_0168>": 64181,
|
181 |
+
"<patch_index_0169>": 64182,
|
182 |
+
"<patch_index_0170>": 64183,
|
183 |
+
"<patch_index_0171>": 64184,
|
184 |
+
"<patch_index_0172>": 64185,
|
185 |
+
"<patch_index_0173>": 64186,
|
186 |
+
"<patch_index_0174>": 64187,
|
187 |
+
"<patch_index_0175>": 64188,
|
188 |
+
"<patch_index_0176>": 64189,
|
189 |
+
"<patch_index_0177>": 64190,
|
190 |
+
"<patch_index_0178>": 64191,
|
191 |
+
"<patch_index_0179>": 64192,
|
192 |
+
"<patch_index_0180>": 64193,
|
193 |
+
"<patch_index_0181>": 64194,
|
194 |
+
"<patch_index_0182>": 64195,
|
195 |
+
"<patch_index_0183>": 64196,
|
196 |
+
"<patch_index_0184>": 64197,
|
197 |
+
"<patch_index_0185>": 64198,
|
198 |
+
"<patch_index_0186>": 64199,
|
199 |
+
"<patch_index_0187>": 64200,
|
200 |
+
"<patch_index_0188>": 64201,
|
201 |
+
"<patch_index_0189>": 64202,
|
202 |
+
"<patch_index_0190>": 64203,
|
203 |
+
"<patch_index_0191>": 64204,
|
204 |
+
"<patch_index_0192>": 64205,
|
205 |
+
"<patch_index_0193>": 64206,
|
206 |
+
"<patch_index_0194>": 64207,
|
207 |
+
"<patch_index_0195>": 64208,
|
208 |
+
"<patch_index_0196>": 64209,
|
209 |
+
"<patch_index_0197>": 64210,
|
210 |
+
"<patch_index_0198>": 64211,
|
211 |
+
"<patch_index_0199>": 64212,
|
212 |
+
"<patch_index_0200>": 64213,
|
213 |
+
"<patch_index_0201>": 64214,
|
214 |
+
"<patch_index_0202>": 64215,
|
215 |
+
"<patch_index_0203>": 64216,
|
216 |
+
"<patch_index_0204>": 64217,
|
217 |
+
"<patch_index_0205>": 64218,
|
218 |
+
"<patch_index_0206>": 64219,
|
219 |
+
"<patch_index_0207>": 64220,
|
220 |
+
"<patch_index_0208>": 64221,
|
221 |
+
"<patch_index_0209>": 64222,
|
222 |
+
"<patch_index_0210>": 64223,
|
223 |
+
"<patch_index_0211>": 64224,
|
224 |
+
"<patch_index_0212>": 64225,
|
225 |
+
"<patch_index_0213>": 64226,
|
226 |
+
"<patch_index_0214>": 64227,
|
227 |
+
"<patch_index_0215>": 64228,
|
228 |
+
"<patch_index_0216>": 64229,
|
229 |
+
"<patch_index_0217>": 64230,
|
230 |
+
"<patch_index_0218>": 64231,
|
231 |
+
"<patch_index_0219>": 64232,
|
232 |
+
"<patch_index_0220>": 64233,
|
233 |
+
"<patch_index_0221>": 64234,
|
234 |
+
"<patch_index_0222>": 64235,
|
235 |
+
"<patch_index_0223>": 64236,
|
236 |
+
"<patch_index_0224>": 64237,
|
237 |
+
"<patch_index_0225>": 64238,
|
238 |
+
"<patch_index_0226>": 64239,
|
239 |
+
"<patch_index_0227>": 64240,
|
240 |
+
"<patch_index_0228>": 64241,
|
241 |
+
"<patch_index_0229>": 64242,
|
242 |
+
"<patch_index_0230>": 64243,
|
243 |
+
"<patch_index_0231>": 64244,
|
244 |
+
"<patch_index_0232>": 64245,
|
245 |
+
"<patch_index_0233>": 64246,
|
246 |
+
"<patch_index_0234>": 64247,
|
247 |
+
"<patch_index_0235>": 64248,
|
248 |
+
"<patch_index_0236>": 64249,
|
249 |
+
"<patch_index_0237>": 64250,
|
250 |
+
"<patch_index_0238>": 64251,
|
251 |
+
"<patch_index_0239>": 64252,
|
252 |
+
"<patch_index_0240>": 64253,
|
253 |
+
"<patch_index_0241>": 64254,
|
254 |
+
"<patch_index_0242>": 64255,
|
255 |
+
"<patch_index_0243>": 64256,
|
256 |
+
"<patch_index_0244>": 64257,
|
257 |
+
"<patch_index_0245>": 64258,
|
258 |
+
"<patch_index_0246>": 64259,
|
259 |
+
"<patch_index_0247>": 64260,
|
260 |
+
"<patch_index_0248>": 64261,
|
261 |
+
"<patch_index_0249>": 64262,
|
262 |
+
"<patch_index_0250>": 64263,
|
263 |
+
"<patch_index_0251>": 64264,
|
264 |
+
"<patch_index_0252>": 64265,
|
265 |
+
"<patch_index_0253>": 64266,
|
266 |
+
"<patch_index_0254>": 64267,
|
267 |
+
"<patch_index_0255>": 64268,
|
268 |
+
"<patch_index_0256>": 64269,
|
269 |
+
"<patch_index_0257>": 64270,
|
270 |
+
"<patch_index_0258>": 64271,
|
271 |
+
"<patch_index_0259>": 64272,
|
272 |
+
"<patch_index_0260>": 64273,
|
273 |
+
"<patch_index_0261>": 64274,
|
274 |
+
"<patch_index_0262>": 64275,
|
275 |
+
"<patch_index_0263>": 64276,
|
276 |
+
"<patch_index_0264>": 64277,
|
277 |
+
"<patch_index_0265>": 64278,
|
278 |
+
"<patch_index_0266>": 64279,
|
279 |
+
"<patch_index_0267>": 64280,
|
280 |
+
"<patch_index_0268>": 64281,
|
281 |
+
"<patch_index_0269>": 64282,
|
282 |
+
"<patch_index_0270>": 64283,
|
283 |
+
"<patch_index_0271>": 64284,
|
284 |
+
"<patch_index_0272>": 64285,
|
285 |
+
"<patch_index_0273>": 64286,
|
286 |
+
"<patch_index_0274>": 64287,
|
287 |
+
"<patch_index_0275>": 64288,
|
288 |
+
"<patch_index_0276>": 64289,
|
289 |
+
"<patch_index_0277>": 64290,
|
290 |
+
"<patch_index_0278>": 64291,
|
291 |
+
"<patch_index_0279>": 64292,
|
292 |
+
"<patch_index_0280>": 64293,
|
293 |
+
"<patch_index_0281>": 64294,
|
294 |
+
"<patch_index_0282>": 64295,
|
295 |
+
"<patch_index_0283>": 64296,
|
296 |
+
"<patch_index_0284>": 64297,
|
297 |
+
"<patch_index_0285>": 64298,
|
298 |
+
"<patch_index_0286>": 64299,
|
299 |
+
"<patch_index_0287>": 64300,
|
300 |
+
"<patch_index_0288>": 64301,
|
301 |
+
"<patch_index_0289>": 64302,
|
302 |
+
"<patch_index_0290>": 64303,
|
303 |
+
"<patch_index_0291>": 64304,
|
304 |
+
"<patch_index_0292>": 64305,
|
305 |
+
"<patch_index_0293>": 64306,
|
306 |
+
"<patch_index_0294>": 64307,
|
307 |
+
"<patch_index_0295>": 64308,
|
308 |
+
"<patch_index_0296>": 64309,
|
309 |
+
"<patch_index_0297>": 64310,
|
310 |
+
"<patch_index_0298>": 64311,
|
311 |
+
"<patch_index_0299>": 64312,
|
312 |
+
"<patch_index_0300>": 64313,
|
313 |
+
"<patch_index_0301>": 64314,
|
314 |
+
"<patch_index_0302>": 64315,
|
315 |
+
"<patch_index_0303>": 64316,
|
316 |
+
"<patch_index_0304>": 64317,
|
317 |
+
"<patch_index_0305>": 64318,
|
318 |
+
"<patch_index_0306>": 64319,
|
319 |
+
"<patch_index_0307>": 64320,
|
320 |
+
"<patch_index_0308>": 64321,
|
321 |
+
"<patch_index_0309>": 64322,
|
322 |
+
"<patch_index_0310>": 64323,
|
323 |
+
"<patch_index_0311>": 64324,
|
324 |
+
"<patch_index_0312>": 64325,
|
325 |
+
"<patch_index_0313>": 64326,
|
326 |
+
"<patch_index_0314>": 64327,
|
327 |
+
"<patch_index_0315>": 64328,
|
328 |
+
"<patch_index_0316>": 64329,
|
329 |
+
"<patch_index_0317>": 64330,
|
330 |
+
"<patch_index_0318>": 64331,
|
331 |
+
"<patch_index_0319>": 64332,
|
332 |
+
"<patch_index_0320>": 64333,
|
333 |
+
"<patch_index_0321>": 64334,
|
334 |
+
"<patch_index_0322>": 64335,
|
335 |
+
"<patch_index_0323>": 64336,
|
336 |
+
"<patch_index_0324>": 64337,
|
337 |
+
"<patch_index_0325>": 64338,
|
338 |
+
"<patch_index_0326>": 64339,
|
339 |
+
"<patch_index_0327>": 64340,
|
340 |
+
"<patch_index_0328>": 64341,
|
341 |
+
"<patch_index_0329>": 64342,
|
342 |
+
"<patch_index_0330>": 64343,
|
343 |
+
"<patch_index_0331>": 64344,
|
344 |
+
"<patch_index_0332>": 64345,
|
345 |
+
"<patch_index_0333>": 64346,
|
346 |
+
"<patch_index_0334>": 64347,
|
347 |
+
"<patch_index_0335>": 64348,
|
348 |
+
"<patch_index_0336>": 64349,
|
349 |
+
"<patch_index_0337>": 64350,
|
350 |
+
"<patch_index_0338>": 64351,
|
351 |
+
"<patch_index_0339>": 64352,
|
352 |
+
"<patch_index_0340>": 64353,
|
353 |
+
"<patch_index_0341>": 64354,
|
354 |
+
"<patch_index_0342>": 64355,
|
355 |
+
"<patch_index_0343>": 64356,
|
356 |
+
"<patch_index_0344>": 64357,
|
357 |
+
"<patch_index_0345>": 64358,
|
358 |
+
"<patch_index_0346>": 64359,
|
359 |
+
"<patch_index_0347>": 64360,
|
360 |
+
"<patch_index_0348>": 64361,
|
361 |
+
"<patch_index_0349>": 64362,
|
362 |
+
"<patch_index_0350>": 64363,
|
363 |
+
"<patch_index_0351>": 64364,
|
364 |
+
"<patch_index_0352>": 64365,
|
365 |
+
"<patch_index_0353>": 64366,
|
366 |
+
"<patch_index_0354>": 64367,
|
367 |
+
"<patch_index_0355>": 64368,
|
368 |
+
"<patch_index_0356>": 64369,
|
369 |
+
"<patch_index_0357>": 64370,
|
370 |
+
"<patch_index_0358>": 64371,
|
371 |
+
"<patch_index_0359>": 64372,
|
372 |
+
"<patch_index_0360>": 64373,
|
373 |
+
"<patch_index_0361>": 64374,
|
374 |
+
"<patch_index_0362>": 64375,
|
375 |
+
"<patch_index_0363>": 64376,
|
376 |
+
"<patch_index_0364>": 64377,
|
377 |
+
"<patch_index_0365>": 64378,
|
378 |
+
"<patch_index_0366>": 64379,
|
379 |
+
"<patch_index_0367>": 64380,
|
380 |
+
"<patch_index_0368>": 64381,
|
381 |
+
"<patch_index_0369>": 64382,
|
382 |
+
"<patch_index_0370>": 64383,
|
383 |
+
"<patch_index_0371>": 64384,
|
384 |
+
"<patch_index_0372>": 64385,
|
385 |
+
"<patch_index_0373>": 64386,
|
386 |
+
"<patch_index_0374>": 64387,
|
387 |
+
"<patch_index_0375>": 64388,
|
388 |
+
"<patch_index_0376>": 64389,
|
389 |
+
"<patch_index_0377>": 64390,
|
390 |
+
"<patch_index_0378>": 64391,
|
391 |
+
"<patch_index_0379>": 64392,
|
392 |
+
"<patch_index_0380>": 64393,
|
393 |
+
"<patch_index_0381>": 64394,
|
394 |
+
"<patch_index_0382>": 64395,
|
395 |
+
"<patch_index_0383>": 64396,
|
396 |
+
"<patch_index_0384>": 64397,
|
397 |
+
"<patch_index_0385>": 64398,
|
398 |
+
"<patch_index_0386>": 64399,
|
399 |
+
"<patch_index_0387>": 64400,
|
400 |
+
"<patch_index_0388>": 64401,
|
401 |
+
"<patch_index_0389>": 64402,
|
402 |
+
"<patch_index_0390>": 64403,
|
403 |
+
"<patch_index_0391>": 64404,
|
404 |
+
"<patch_index_0392>": 64405,
|
405 |
+
"<patch_index_0393>": 64406,
|
406 |
+
"<patch_index_0394>": 64407,
|
407 |
+
"<patch_index_0395>": 64408,
|
408 |
+
"<patch_index_0396>": 64409,
|
409 |
+
"<patch_index_0397>": 64410,
|
410 |
+
"<patch_index_0398>": 64411,
|
411 |
+
"<patch_index_0399>": 64412,
|
412 |
+
"<patch_index_0400>": 64413,
|
413 |
+
"<patch_index_0401>": 64414,
|
414 |
+
"<patch_index_0402>": 64415,
|
415 |
+
"<patch_index_0403>": 64416,
|
416 |
+
"<patch_index_0404>": 64417,
|
417 |
+
"<patch_index_0405>": 64418,
|
418 |
+
"<patch_index_0406>": 64419,
|
419 |
+
"<patch_index_0407>": 64420,
|
420 |
+
"<patch_index_0408>": 64421,
|
421 |
+
"<patch_index_0409>": 64422,
|
422 |
+
"<patch_index_0410>": 64423,
|
423 |
+
"<patch_index_0411>": 64424,
|
424 |
+
"<patch_index_0412>": 64425,
|
425 |
+
"<patch_index_0413>": 64426,
|
426 |
+
"<patch_index_0414>": 64427,
|
427 |
+
"<patch_index_0415>": 64428,
|
428 |
+
"<patch_index_0416>": 64429,
|
429 |
+
"<patch_index_0417>": 64430,
|
430 |
+
"<patch_index_0418>": 64431,
|
431 |
+
"<patch_index_0419>": 64432,
|
432 |
+
"<patch_index_0420>": 64433,
|
433 |
+
"<patch_index_0421>": 64434,
|
434 |
+
"<patch_index_0422>": 64435,
|
435 |
+
"<patch_index_0423>": 64436,
|
436 |
+
"<patch_index_0424>": 64437,
|
437 |
+
"<patch_index_0425>": 64438,
|
438 |
+
"<patch_index_0426>": 64439,
|
439 |
+
"<patch_index_0427>": 64440,
|
440 |
+
"<patch_index_0428>": 64441,
|
441 |
+
"<patch_index_0429>": 64442,
|
442 |
+
"<patch_index_0430>": 64443,
|
443 |
+
"<patch_index_0431>": 64444,
|
444 |
+
"<patch_index_0432>": 64445,
|
445 |
+
"<patch_index_0433>": 64446,
|
446 |
+
"<patch_index_0434>": 64447,
|
447 |
+
"<patch_index_0435>": 64448,
|
448 |
+
"<patch_index_0436>": 64449,
|
449 |
+
"<patch_index_0437>": 64450,
|
450 |
+
"<patch_index_0438>": 64451,
|
451 |
+
"<patch_index_0439>": 64452,
|
452 |
+
"<patch_index_0440>": 64453,
|
453 |
+
"<patch_index_0441>": 64454,
|
454 |
+
"<patch_index_0442>": 64455,
|
455 |
+
"<patch_index_0443>": 64456,
|
456 |
+
"<patch_index_0444>": 64457,
|
457 |
+
"<patch_index_0445>": 64458,
|
458 |
+
"<patch_index_0446>": 64459,
|
459 |
+
"<patch_index_0447>": 64460,
|
460 |
+
"<patch_index_0448>": 64461,
|
461 |
+
"<patch_index_0449>": 64462,
|
462 |
+
"<patch_index_0450>": 64463,
|
463 |
+
"<patch_index_0451>": 64464,
|
464 |
+
"<patch_index_0452>": 64465,
|
465 |
+
"<patch_index_0453>": 64466,
|
466 |
+
"<patch_index_0454>": 64467,
|
467 |
+
"<patch_index_0455>": 64468,
|
468 |
+
"<patch_index_0456>": 64469,
|
469 |
+
"<patch_index_0457>": 64470,
|
470 |
+
"<patch_index_0458>": 64471,
|
471 |
+
"<patch_index_0459>": 64472,
|
472 |
+
"<patch_index_0460>": 64473,
|
473 |
+
"<patch_index_0461>": 64474,
|
474 |
+
"<patch_index_0462>": 64475,
|
475 |
+
"<patch_index_0463>": 64476,
|
476 |
+
"<patch_index_0464>": 64477,
|
477 |
+
"<patch_index_0465>": 64478,
|
478 |
+
"<patch_index_0466>": 64479,
|
479 |
+
"<patch_index_0467>": 64480,
|
480 |
+
"<patch_index_0468>": 64481,
|
481 |
+
"<patch_index_0469>": 64482,
|
482 |
+
"<patch_index_0470>": 64483,
|
483 |
+
"<patch_index_0471>": 64484,
|
484 |
+
"<patch_index_0472>": 64485,
|
485 |
+
"<patch_index_0473>": 64486,
|
486 |
+
"<patch_index_0474>": 64487,
|
487 |
+
"<patch_index_0475>": 64488,
|
488 |
+
"<patch_index_0476>": 64489,
|
489 |
+
"<patch_index_0477>": 64490,
|
490 |
+
"<patch_index_0478>": 64491,
|
491 |
+
"<patch_index_0479>": 64492,
|
492 |
+
"<patch_index_0480>": 64493,
|
493 |
+
"<patch_index_0481>": 64494,
|
494 |
+
"<patch_index_0482>": 64495,
|
495 |
+
"<patch_index_0483>": 64496,
|
496 |
+
"<patch_index_0484>": 64497,
|
497 |
+
"<patch_index_0485>": 64498,
|
498 |
+
"<patch_index_0486>": 64499,
|
499 |
+
"<patch_index_0487>": 64500,
|
500 |
+
"<patch_index_0488>": 64501,
|
501 |
+
"<patch_index_0489>": 64502,
|
502 |
+
"<patch_index_0490>": 64503,
|
503 |
+
"<patch_index_0491>": 64504,
|
504 |
+
"<patch_index_0492>": 64505,
|
505 |
+
"<patch_index_0493>": 64506,
|
506 |
+
"<patch_index_0494>": 64507,
|
507 |
+
"<patch_index_0495>": 64508,
|
508 |
+
"<patch_index_0496>": 64509,
|
509 |
+
"<patch_index_0497>": 64510,
|
510 |
+
"<patch_index_0498>": 64511,
|
511 |
+
"<patch_index_0499>": 64512,
|
512 |
+
"<patch_index_0500>": 64513,
|
513 |
+
"<patch_index_0501>": 64514,
|
514 |
+
"<patch_index_0502>": 64515,
|
515 |
+
"<patch_index_0503>": 64516,
|
516 |
+
"<patch_index_0504>": 64517,
|
517 |
+
"<patch_index_0505>": 64518,
|
518 |
+
"<patch_index_0506>": 64519,
|
519 |
+
"<patch_index_0507>": 64520,
|
520 |
+
"<patch_index_0508>": 64521,
|
521 |
+
"<patch_index_0509>": 64522,
|
522 |
+
"<patch_index_0510>": 64523,
|
523 |
+
"<patch_index_0511>": 64524,
|
524 |
+
"<patch_index_0512>": 64525,
|
525 |
+
"<patch_index_0513>": 64526,
|
526 |
+
"<patch_index_0514>": 64527,
|
527 |
+
"<patch_index_0515>": 64528,
|
528 |
+
"<patch_index_0516>": 64529,
|
529 |
+
"<patch_index_0517>": 64530,
|
530 |
+
"<patch_index_0518>": 64531,
|
531 |
+
"<patch_index_0519>": 64532,
|
532 |
+
"<patch_index_0520>": 64533,
|
533 |
+
"<patch_index_0521>": 64534,
|
534 |
+
"<patch_index_0522>": 64535,
|
535 |
+
"<patch_index_0523>": 64536,
|
536 |
+
"<patch_index_0524>": 64537,
|
537 |
+
"<patch_index_0525>": 64538,
|
538 |
+
"<patch_index_0526>": 64539,
|
539 |
+
"<patch_index_0527>": 64540,
|
540 |
+
"<patch_index_0528>": 64541,
|
541 |
+
"<patch_index_0529>": 64542,
|
542 |
+
"<patch_index_0530>": 64543,
|
543 |
+
"<patch_index_0531>": 64544,
|
544 |
+
"<patch_index_0532>": 64545,
|
545 |
+
"<patch_index_0533>": 64546,
|
546 |
+
"<patch_index_0534>": 64547,
|
547 |
+
"<patch_index_0535>": 64548,
|
548 |
+
"<patch_index_0536>": 64549,
|
549 |
+
"<patch_index_0537>": 64550,
|
550 |
+
"<patch_index_0538>": 64551,
|
551 |
+
"<patch_index_0539>": 64552,
|
552 |
+
"<patch_index_0540>": 64553,
|
553 |
+
"<patch_index_0541>": 64554,
|
554 |
+
"<patch_index_0542>": 64555,
|
555 |
+
"<patch_index_0543>": 64556,
|
556 |
+
"<patch_index_0544>": 64557,
|
557 |
+
"<patch_index_0545>": 64558,
|
558 |
+
"<patch_index_0546>": 64559,
|
559 |
+
"<patch_index_0547>": 64560,
|
560 |
+
"<patch_index_0548>": 64561,
|
561 |
+
"<patch_index_0549>": 64562,
|
562 |
+
"<patch_index_0550>": 64563,
|
563 |
+
"<patch_index_0551>": 64564,
|
564 |
+
"<patch_index_0552>": 64565,
|
565 |
+
"<patch_index_0553>": 64566,
|
566 |
+
"<patch_index_0554>": 64567,
|
567 |
+
"<patch_index_0555>": 64568,
|
568 |
+
"<patch_index_0556>": 64569,
|
569 |
+
"<patch_index_0557>": 64570,
|
570 |
+
"<patch_index_0558>": 64571,
|
571 |
+
"<patch_index_0559>": 64572,
|
572 |
+
"<patch_index_0560>": 64573,
|
573 |
+
"<patch_index_0561>": 64574,
|
574 |
+
"<patch_index_0562>": 64575,
|
575 |
+
"<patch_index_0563>": 64576,
|
576 |
+
"<patch_index_0564>": 64577,
|
577 |
+
"<patch_index_0565>": 64578,
|
578 |
+
"<patch_index_0566>": 64579,
|
579 |
+
"<patch_index_0567>": 64580,
|
580 |
+
"<patch_index_0568>": 64581,
|
581 |
+
"<patch_index_0569>": 64582,
|
582 |
+
"<patch_index_0570>": 64583,
|
583 |
+
"<patch_index_0571>": 64584,
|
584 |
+
"<patch_index_0572>": 64585,
|
585 |
+
"<patch_index_0573>": 64586,
|
586 |
+
"<patch_index_0574>": 64587,
|
587 |
+
"<patch_index_0575>": 64588,
|
588 |
+
"<patch_index_0576>": 64589,
|
589 |
+
"<patch_index_0577>": 64590,
|
590 |
+
"<patch_index_0578>": 64591,
|
591 |
+
"<patch_index_0579>": 64592,
|
592 |
+
"<patch_index_0580>": 64593,
|
593 |
+
"<patch_index_0581>": 64594,
|
594 |
+
"<patch_index_0582>": 64595,
|
595 |
+
"<patch_index_0583>": 64596,
|
596 |
+
"<patch_index_0584>": 64597,
|
597 |
+
"<patch_index_0585>": 64598,
|
598 |
+
"<patch_index_0586>": 64599,
|
599 |
+
"<patch_index_0587>": 64600,
|
600 |
+
"<patch_index_0588>": 64601,
|
601 |
+
"<patch_index_0589>": 64602,
|
602 |
+
"<patch_index_0590>": 64603,
|
603 |
+
"<patch_index_0591>": 64604,
|
604 |
+
"<patch_index_0592>": 64605,
|
605 |
+
"<patch_index_0593>": 64606,
|
606 |
+
"<patch_index_0594>": 64607,
|
607 |
+
"<patch_index_0595>": 64608,
|
608 |
+
"<patch_index_0596>": 64609,
|
609 |
+
"<patch_index_0597>": 64610,
|
610 |
+
"<patch_index_0598>": 64611,
|
611 |
+
"<patch_index_0599>": 64612,
|
612 |
+
"<patch_index_0600>": 64613,
|
613 |
+
"<patch_index_0601>": 64614,
|
614 |
+
"<patch_index_0602>": 64615,
|
615 |
+
"<patch_index_0603>": 64616,
|
616 |
+
"<patch_index_0604>": 64617,
|
617 |
+
"<patch_index_0605>": 64618,
|
618 |
+
"<patch_index_0606>": 64619,
|
619 |
+
"<patch_index_0607>": 64620,
|
620 |
+
"<patch_index_0608>": 64621,
|
621 |
+
"<patch_index_0609>": 64622,
|
622 |
+
"<patch_index_0610>": 64623,
|
623 |
+
"<patch_index_0611>": 64624,
|
624 |
+
"<patch_index_0612>": 64625,
|
625 |
+
"<patch_index_0613>": 64626,
|
626 |
+
"<patch_index_0614>": 64627,
|
627 |
+
"<patch_index_0615>": 64628,
|
628 |
+
"<patch_index_0616>": 64629,
|
629 |
+
"<patch_index_0617>": 64630,
|
630 |
+
"<patch_index_0618>": 64631,
|
631 |
+
"<patch_index_0619>": 64632,
|
632 |
+
"<patch_index_0620>": 64633,
|
633 |
+
"<patch_index_0621>": 64634,
|
634 |
+
"<patch_index_0622>": 64635,
|
635 |
+
"<patch_index_0623>": 64636,
|
636 |
+
"<patch_index_0624>": 64637,
|
637 |
+
"<patch_index_0625>": 64638,
|
638 |
+
"<patch_index_0626>": 64639,
|
639 |
+
"<patch_index_0627>": 64640,
|
640 |
+
"<patch_index_0628>": 64641,
|
641 |
+
"<patch_index_0629>": 64642,
|
642 |
+
"<patch_index_0630>": 64643,
|
643 |
+
"<patch_index_0631>": 64644,
|
644 |
+
"<patch_index_0632>": 64645,
|
645 |
+
"<patch_index_0633>": 64646,
|
646 |
+
"<patch_index_0634>": 64647,
|
647 |
+
"<patch_index_0635>": 64648,
|
648 |
+
"<patch_index_0636>": 64649,
|
649 |
+
"<patch_index_0637>": 64650,
|
650 |
+
"<patch_index_0638>": 64651,
|
651 |
+
"<patch_index_0639>": 64652,
|
652 |
+
"<patch_index_0640>": 64653,
|
653 |
+
"<patch_index_0641>": 64654,
|
654 |
+
"<patch_index_0642>": 64655,
|
655 |
+
"<patch_index_0643>": 64656,
|
656 |
+
"<patch_index_0644>": 64657,
|
657 |
+
"<patch_index_0645>": 64658,
|
658 |
+
"<patch_index_0646>": 64659,
|
659 |
+
"<patch_index_0647>": 64660,
|
660 |
+
"<patch_index_0648>": 64661,
|
661 |
+
"<patch_index_0649>": 64662,
|
662 |
+
"<patch_index_0650>": 64663,
|
663 |
+
"<patch_index_0651>": 64664,
|
664 |
+
"<patch_index_0652>": 64665,
|
665 |
+
"<patch_index_0653>": 64666,
|
666 |
+
"<patch_index_0654>": 64667,
|
667 |
+
"<patch_index_0655>": 64668,
|
668 |
+
"<patch_index_0656>": 64669,
|
669 |
+
"<patch_index_0657>": 64670,
|
670 |
+
"<patch_index_0658>": 64671,
|
671 |
+
"<patch_index_0659>": 64672,
|
672 |
+
"<patch_index_0660>": 64673,
|
673 |
+
"<patch_index_0661>": 64674,
|
674 |
+
"<patch_index_0662>": 64675,
|
675 |
+
"<patch_index_0663>": 64676,
|
676 |
+
"<patch_index_0664>": 64677,
|
677 |
+
"<patch_index_0665>": 64678,
|
678 |
+
"<patch_index_0666>": 64679,
|
679 |
+
"<patch_index_0667>": 64680,
|
680 |
+
"<patch_index_0668>": 64681,
|
681 |
+
"<patch_index_0669>": 64682,
|
682 |
+
"<patch_index_0670>": 64683,
|
683 |
+
"<patch_index_0671>": 64684,
|
684 |
+
"<patch_index_0672>": 64685,
|
685 |
+
"<patch_index_0673>": 64686,
|
686 |
+
"<patch_index_0674>": 64687,
|
687 |
+
"<patch_index_0675>": 64688,
|
688 |
+
"<patch_index_0676>": 64689,
|
689 |
+
"<patch_index_0677>": 64690,
|
690 |
+
"<patch_index_0678>": 64691,
|
691 |
+
"<patch_index_0679>": 64692,
|
692 |
+
"<patch_index_0680>": 64693,
|
693 |
+
"<patch_index_0681>": 64694,
|
694 |
+
"<patch_index_0682>": 64695,
|
695 |
+
"<patch_index_0683>": 64696,
|
696 |
+
"<patch_index_0684>": 64697,
|
697 |
+
"<patch_index_0685>": 64698,
|
698 |
+
"<patch_index_0686>": 64699,
|
699 |
+
"<patch_index_0687>": 64700,
|
700 |
+
"<patch_index_0688>": 64701,
|
701 |
+
"<patch_index_0689>": 64702,
|
702 |
+
"<patch_index_0690>": 64703,
|
703 |
+
"<patch_index_0691>": 64704,
|
704 |
+
"<patch_index_0692>": 64705,
|
705 |
+
"<patch_index_0693>": 64706,
|
706 |
+
"<patch_index_0694>": 64707,
|
707 |
+
"<patch_index_0695>": 64708,
|
708 |
+
"<patch_index_0696>": 64709,
|
709 |
+
"<patch_index_0697>": 64710,
|
710 |
+
"<patch_index_0698>": 64711,
|
711 |
+
"<patch_index_0699>": 64712,
|
712 |
+
"<patch_index_0700>": 64713,
|
713 |
+
"<patch_index_0701>": 64714,
|
714 |
+
"<patch_index_0702>": 64715,
|
715 |
+
"<patch_index_0703>": 64716,
|
716 |
+
"<patch_index_0704>": 64717,
|
717 |
+
"<patch_index_0705>": 64718,
|
718 |
+
"<patch_index_0706>": 64719,
|
719 |
+
"<patch_index_0707>": 64720,
|
720 |
+
"<patch_index_0708>": 64721,
|
721 |
+
"<patch_index_0709>": 64722,
|
722 |
+
"<patch_index_0710>": 64723,
|
723 |
+
"<patch_index_0711>": 64724,
|
724 |
+
"<patch_index_0712>": 64725,
|
725 |
+
"<patch_index_0713>": 64726,
|
726 |
+
"<patch_index_0714>": 64727,
|
727 |
+
"<patch_index_0715>": 64728,
|
728 |
+
"<patch_index_0716>": 64729,
|
729 |
+
"<patch_index_0717>": 64730,
|
730 |
+
"<patch_index_0718>": 64731,
|
731 |
+
"<patch_index_0719>": 64732,
|
732 |
+
"<patch_index_0720>": 64733,
|
733 |
+
"<patch_index_0721>": 64734,
|
734 |
+
"<patch_index_0722>": 64735,
|
735 |
+
"<patch_index_0723>": 64736,
|
736 |
+
"<patch_index_0724>": 64737,
|
737 |
+
"<patch_index_0725>": 64738,
|
738 |
+
"<patch_index_0726>": 64739,
|
739 |
+
"<patch_index_0727>": 64740,
|
740 |
+
"<patch_index_0728>": 64741,
|
741 |
+
"<patch_index_0729>": 64742,
|
742 |
+
"<patch_index_0730>": 64743,
|
743 |
+
"<patch_index_0731>": 64744,
|
744 |
+
"<patch_index_0732>": 64745,
|
745 |
+
"<patch_index_0733>": 64746,
|
746 |
+
"<patch_index_0734>": 64747,
|
747 |
+
"<patch_index_0735>": 64748,
|
748 |
+
"<patch_index_0736>": 64749,
|
749 |
+
"<patch_index_0737>": 64750,
|
750 |
+
"<patch_index_0738>": 64751,
|
751 |
+
"<patch_index_0739>": 64752,
|
752 |
+
"<patch_index_0740>": 64753,
|
753 |
+
"<patch_index_0741>": 64754,
|
754 |
+
"<patch_index_0742>": 64755,
|
755 |
+
"<patch_index_0743>": 64756,
|
756 |
+
"<patch_index_0744>": 64757,
|
757 |
+
"<patch_index_0745>": 64758,
|
758 |
+
"<patch_index_0746>": 64759,
|
759 |
+
"<patch_index_0747>": 64760,
|
760 |
+
"<patch_index_0748>": 64761,
|
761 |
+
"<patch_index_0749>": 64762,
|
762 |
+
"<patch_index_0750>": 64763,
|
763 |
+
"<patch_index_0751>": 64764,
|
764 |
+
"<patch_index_0752>": 64765,
|
765 |
+
"<patch_index_0753>": 64766,
|
766 |
+
"<patch_index_0754>": 64767,
|
767 |
+
"<patch_index_0755>": 64768,
|
768 |
+
"<patch_index_0756>": 64769,
|
769 |
+
"<patch_index_0757>": 64770,
|
770 |
+
"<patch_index_0758>": 64771,
|
771 |
+
"<patch_index_0759>": 64772,
|
772 |
+
"<patch_index_0760>": 64773,
|
773 |
+
"<patch_index_0761>": 64774,
|
774 |
+
"<patch_index_0762>": 64775,
|
775 |
+
"<patch_index_0763>": 64776,
|
776 |
+
"<patch_index_0764>": 64777,
|
777 |
+
"<patch_index_0765>": 64778,
|
778 |
+
"<patch_index_0766>": 64779,
|
779 |
+
"<patch_index_0767>": 64780,
|
780 |
+
"<patch_index_0768>": 64781,
|
781 |
+
"<patch_index_0769>": 64782,
|
782 |
+
"<patch_index_0770>": 64783,
|
783 |
+
"<patch_index_0771>": 64784,
|
784 |
+
"<patch_index_0772>": 64785,
|
785 |
+
"<patch_index_0773>": 64786,
|
786 |
+
"<patch_index_0774>": 64787,
|
787 |
+
"<patch_index_0775>": 64788,
|
788 |
+
"<patch_index_0776>": 64789,
|
789 |
+
"<patch_index_0777>": 64790,
|
790 |
+
"<patch_index_0778>": 64791,
|
791 |
+
"<patch_index_0779>": 64792,
|
792 |
+
"<patch_index_0780>": 64793,
|
793 |
+
"<patch_index_0781>": 64794,
|
794 |
+
"<patch_index_0782>": 64795,
|
795 |
+
"<patch_index_0783>": 64796,
|
796 |
+
"<patch_index_0784>": 64797,
|
797 |
+
"<patch_index_0785>": 64798,
|
798 |
+
"<patch_index_0786>": 64799,
|
799 |
+
"<patch_index_0787>": 64800,
|
800 |
+
"<patch_index_0788>": 64801,
|
801 |
+
"<patch_index_0789>": 64802,
|
802 |
+
"<patch_index_0790>": 64803,
|
803 |
+
"<patch_index_0791>": 64804,
|
804 |
+
"<patch_index_0792>": 64805,
|
805 |
+
"<patch_index_0793>": 64806,
|
806 |
+
"<patch_index_0794>": 64807,
|
807 |
+
"<patch_index_0795>": 64808,
|
808 |
+
"<patch_index_0796>": 64809,
|
809 |
+
"<patch_index_0797>": 64810,
|
810 |
+
"<patch_index_0798>": 64811,
|
811 |
+
"<patch_index_0799>": 64812,
|
812 |
+
"<patch_index_0800>": 64813,
|
813 |
+
"<patch_index_0801>": 64814,
|
814 |
+
"<patch_index_0802>": 64815,
|
815 |
+
"<patch_index_0803>": 64816,
|
816 |
+
"<patch_index_0804>": 64817,
|
817 |
+
"<patch_index_0805>": 64818,
|
818 |
+
"<patch_index_0806>": 64819,
|
819 |
+
"<patch_index_0807>": 64820,
|
820 |
+
"<patch_index_0808>": 64821,
|
821 |
+
"<patch_index_0809>": 64822,
|
822 |
+
"<patch_index_0810>": 64823,
|
823 |
+
"<patch_index_0811>": 64824,
|
824 |
+
"<patch_index_0812>": 64825,
|
825 |
+
"<patch_index_0813>": 64826,
|
826 |
+
"<patch_index_0814>": 64827,
|
827 |
+
"<patch_index_0815>": 64828,
|
828 |
+
"<patch_index_0816>": 64829,
|
829 |
+
"<patch_index_0817>": 64830,
|
830 |
+
"<patch_index_0818>": 64831,
|
831 |
+
"<patch_index_0819>": 64832,
|
832 |
+
"<patch_index_0820>": 64833,
|
833 |
+
"<patch_index_0821>": 64834,
|
834 |
+
"<patch_index_0822>": 64835,
|
835 |
+
"<patch_index_0823>": 64836,
|
836 |
+
"<patch_index_0824>": 64837,
|
837 |
+
"<patch_index_0825>": 64838,
|
838 |
+
"<patch_index_0826>": 64839,
|
839 |
+
"<patch_index_0827>": 64840,
|
840 |
+
"<patch_index_0828>": 64841,
|
841 |
+
"<patch_index_0829>": 64842,
|
842 |
+
"<patch_index_0830>": 64843,
|
843 |
+
"<patch_index_0831>": 64844,
|
844 |
+
"<patch_index_0832>": 64845,
|
845 |
+
"<patch_index_0833>": 64846,
|
846 |
+
"<patch_index_0834>": 64847,
|
847 |
+
"<patch_index_0835>": 64848,
|
848 |
+
"<patch_index_0836>": 64849,
|
849 |
+
"<patch_index_0837>": 64850,
|
850 |
+
"<patch_index_0838>": 64851,
|
851 |
+
"<patch_index_0839>": 64852,
|
852 |
+
"<patch_index_0840>": 64853,
|
853 |
+
"<patch_index_0841>": 64854,
|
854 |
+
"<patch_index_0842>": 64855,
|
855 |
+
"<patch_index_0843>": 64856,
|
856 |
+
"<patch_index_0844>": 64857,
|
857 |
+
"<patch_index_0845>": 64858,
|
858 |
+
"<patch_index_0846>": 64859,
|
859 |
+
"<patch_index_0847>": 64860,
|
860 |
+
"<patch_index_0848>": 64861,
|
861 |
+
"<patch_index_0849>": 64862,
|
862 |
+
"<patch_index_0850>": 64863,
|
863 |
+
"<patch_index_0851>": 64864,
|
864 |
+
"<patch_index_0852>": 64865,
|
865 |
+
"<patch_index_0853>": 64866,
|
866 |
+
"<patch_index_0854>": 64867,
|
867 |
+
"<patch_index_0855>": 64868,
|
868 |
+
"<patch_index_0856>": 64869,
|
869 |
+
"<patch_index_0857>": 64870,
|
870 |
+
"<patch_index_0858>": 64871,
|
871 |
+
"<patch_index_0859>": 64872,
|
872 |
+
"<patch_index_0860>": 64873,
|
873 |
+
"<patch_index_0861>": 64874,
|
874 |
+
"<patch_index_0862>": 64875,
|
875 |
+
"<patch_index_0863>": 64876,
|
876 |
+
"<patch_index_0864>": 64877,
|
877 |
+
"<patch_index_0865>": 64878,
|
878 |
+
"<patch_index_0866>": 64879,
|
879 |
+
"<patch_index_0867>": 64880,
|
880 |
+
"<patch_index_0868>": 64881,
|
881 |
+
"<patch_index_0869>": 64882,
|
882 |
+
"<patch_index_0870>": 64883,
|
883 |
+
"<patch_index_0871>": 64884,
|
884 |
+
"<patch_index_0872>": 64885,
|
885 |
+
"<patch_index_0873>": 64886,
|
886 |
+
"<patch_index_0874>": 64887,
|
887 |
+
"<patch_index_0875>": 64888,
|
888 |
+
"<patch_index_0876>": 64889,
|
889 |
+
"<patch_index_0877>": 64890,
|
890 |
+
"<patch_index_0878>": 64891,
|
891 |
+
"<patch_index_0879>": 64892,
|
892 |
+
"<patch_index_0880>": 64893,
|
893 |
+
"<patch_index_0881>": 64894,
|
894 |
+
"<patch_index_0882>": 64895,
|
895 |
+
"<patch_index_0883>": 64896,
|
896 |
+
"<patch_index_0884>": 64897,
|
897 |
+
"<patch_index_0885>": 64898,
|
898 |
+
"<patch_index_0886>": 64899,
|
899 |
+
"<patch_index_0887>": 64900,
|
900 |
+
"<patch_index_0888>": 64901,
|
901 |
+
"<patch_index_0889>": 64902,
|
902 |
+
"<patch_index_0890>": 64903,
|
903 |
+
"<patch_index_0891>": 64904,
|
904 |
+
"<patch_index_0892>": 64905,
|
905 |
+
"<patch_index_0893>": 64906,
|
906 |
+
"<patch_index_0894>": 64907,
|
907 |
+
"<patch_index_0895>": 64908,
|
908 |
+
"<patch_index_0896>": 64909,
|
909 |
+
"<patch_index_0897>": 64910,
|
910 |
+
"<patch_index_0898>": 64911,
|
911 |
+
"<patch_index_0899>": 64912,
|
912 |
+
"<patch_index_0900>": 64913,
|
913 |
+
"<patch_index_0901>": 64914,
|
914 |
+
"<patch_index_0902>": 64915,
|
915 |
+
"<patch_index_0903>": 64916,
|
916 |
+
"<patch_index_0904>": 64917,
|
917 |
+
"<patch_index_0905>": 64918,
|
918 |
+
"<patch_index_0906>": 64919,
|
919 |
+
"<patch_index_0907>": 64920,
|
920 |
+
"<patch_index_0908>": 64921,
|
921 |
+
"<patch_index_0909>": 64922,
|
922 |
+
"<patch_index_0910>": 64923,
|
923 |
+
"<patch_index_0911>": 64924,
|
924 |
+
"<patch_index_0912>": 64925,
|
925 |
+
"<patch_index_0913>": 64926,
|
926 |
+
"<patch_index_0914>": 64927,
|
927 |
+
"<patch_index_0915>": 64928,
|
928 |
+
"<patch_index_0916>": 64929,
|
929 |
+
"<patch_index_0917>": 64930,
|
930 |
+
"<patch_index_0918>": 64931,
|
931 |
+
"<patch_index_0919>": 64932,
|
932 |
+
"<patch_index_0920>": 64933,
|
933 |
+
"<patch_index_0921>": 64934,
|
934 |
+
"<patch_index_0922>": 64935,
|
935 |
+
"<patch_index_0923>": 64936,
|
936 |
+
"<patch_index_0924>": 64937,
|
937 |
+
"<patch_index_0925>": 64938,
|
938 |
+
"<patch_index_0926>": 64939,
|
939 |
+
"<patch_index_0927>": 64940,
|
940 |
+
"<patch_index_0928>": 64941,
|
941 |
+
"<patch_index_0929>": 64942,
|
942 |
+
"<patch_index_0930>": 64943,
|
943 |
+
"<patch_index_0931>": 64944,
|
944 |
+
"<patch_index_0932>": 64945,
|
945 |
+
"<patch_index_0933>": 64946,
|
946 |
+
"<patch_index_0934>": 64947,
|
947 |
+
"<patch_index_0935>": 64948,
|
948 |
+
"<patch_index_0936>": 64949,
|
949 |
+
"<patch_index_0937>": 64950,
|
950 |
+
"<patch_index_0938>": 64951,
|
951 |
+
"<patch_index_0939>": 64952,
|
952 |
+
"<patch_index_0940>": 64953,
|
953 |
+
"<patch_index_0941>": 64954,
|
954 |
+
"<patch_index_0942>": 64955,
|
955 |
+
"<patch_index_0943>": 64956,
|
956 |
+
"<patch_index_0944>": 64957,
|
957 |
+
"<patch_index_0945>": 64958,
|
958 |
+
"<patch_index_0946>": 64959,
|
959 |
+
"<patch_index_0947>": 64960,
|
960 |
+
"<patch_index_0948>": 64961,
|
961 |
+
"<patch_index_0949>": 64962,
|
962 |
+
"<patch_index_0950>": 64963,
|
963 |
+
"<patch_index_0951>": 64964,
|
964 |
+
"<patch_index_0952>": 64965,
|
965 |
+
"<patch_index_0953>": 64966,
|
966 |
+
"<patch_index_0954>": 64967,
|
967 |
+
"<patch_index_0955>": 64968,
|
968 |
+
"<patch_index_0956>": 64969,
|
969 |
+
"<patch_index_0957>": 64970,
|
970 |
+
"<patch_index_0958>": 64971,
|
971 |
+
"<patch_index_0959>": 64972,
|
972 |
+
"<patch_index_0960>": 64973,
|
973 |
+
"<patch_index_0961>": 64974,
|
974 |
+
"<patch_index_0962>": 64975,
|
975 |
+
"<patch_index_0963>": 64976,
|
976 |
+
"<patch_index_0964>": 64977,
|
977 |
+
"<patch_index_0965>": 64978,
|
978 |
+
"<patch_index_0966>": 64979,
|
979 |
+
"<patch_index_0967>": 64980,
|
980 |
+
"<patch_index_0968>": 64981,
|
981 |
+
"<patch_index_0969>": 64982,
|
982 |
+
"<patch_index_0970>": 64983,
|
983 |
+
"<patch_index_0971>": 64984,
|
984 |
+
"<patch_index_0972>": 64985,
|
985 |
+
"<patch_index_0973>": 64986,
|
986 |
+
"<patch_index_0974>": 64987,
|
987 |
+
"<patch_index_0975>": 64988,
|
988 |
+
"<patch_index_0976>": 64989,
|
989 |
+
"<patch_index_0977>": 64990,
|
990 |
+
"<patch_index_0978>": 64991,
|
991 |
+
"<patch_index_0979>": 64992,
|
992 |
+
"<patch_index_0980>": 64993,
|
993 |
+
"<patch_index_0981>": 64994,
|
994 |
+
"<patch_index_0982>": 64995,
|
995 |
+
"<patch_index_0983>": 64996,
|
996 |
+
"<patch_index_0984>": 64997,
|
997 |
+
"<patch_index_0985>": 64998,
|
998 |
+
"<patch_index_0986>": 64999,
|
999 |
+
"<patch_index_0987>": 65000,
|
1000 |
+
"<patch_index_0988>": 65001,
|
1001 |
+
"<patch_index_0989>": 65002,
|
1002 |
+
"<patch_index_0990>": 65003,
|
1003 |
+
"<patch_index_0991>": 65004,
|
1004 |
+
"<patch_index_0992>": 65005,
|
1005 |
+
"<patch_index_0993>": 65006,
|
1006 |
+
"<patch_index_0994>": 65007,
|
1007 |
+
"<patch_index_0995>": 65008,
|
1008 |
+
"<patch_index_0996>": 65009,
|
1009 |
+
"<patch_index_0997>": 65010,
|
1010 |
+
"<patch_index_0998>": 65011,
|
1011 |
+
"<patch_index_0999>": 65012,
|
1012 |
+
"<patch_index_1000>": 65013,
|
1013 |
+
"<patch_index_1001>": 65014,
|
1014 |
+
"<patch_index_1002>": 65015,
|
1015 |
+
"<patch_index_1003>": 65016,
|
1016 |
+
"<patch_index_1004>": 65017,
|
1017 |
+
"<patch_index_1005>": 65018,
|
1018 |
+
"<patch_index_1006>": 65019,
|
1019 |
+
"<patch_index_1007>": 65020,
|
1020 |
+
"<patch_index_1008>": 65021,
|
1021 |
+
"<patch_index_1009>": 65022,
|
1022 |
+
"<patch_index_1010>": 65023,
|
1023 |
+
"<patch_index_1011>": 65024,
|
1024 |
+
"<patch_index_1012>": 65025,
|
1025 |
+
"<patch_index_1013>": 65026,
|
1026 |
+
"<patch_index_1014>": 65027,
|
1027 |
+
"<patch_index_1015>": 65028,
|
1028 |
+
"<patch_index_1016>": 65029,
|
1029 |
+
"<patch_index_1017>": 65030,
|
1030 |
+
"<patch_index_1018>": 65031,
|
1031 |
+
"<patch_index_1019>": 65032,
|
1032 |
+
"<patch_index_1020>": 65033,
|
1033 |
+
"<patch_index_1021>": 65034,
|
1034 |
+
"<patch_index_1022>": 65035,
|
1035 |
+
"<patch_index_1023>": 65036,
|
1036 |
+
"<phrase>": 64007
|
1037 |
+
}
|
config.json
ADDED
@@ -0,0 +1,163 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/kosmos-2-patch14-224",
|
3 |
+
"architectures": [
|
4 |
+
"Kosmos2ForConditionalGeneration"
|
5 |
+
],
|
6 |
+
"latent_query_num": 64,
|
7 |
+
"model_type": "kosmos-2",
|
8 |
+
"text_config": {
|
9 |
+
"_name_or_path": "",
|
10 |
+
"activation_dropout": 0.0,
|
11 |
+
"activation_function": "gelu",
|
12 |
+
"add_cross_attention": false,
|
13 |
+
"architectures": null,
|
14 |
+
"attention_dropout": 0.1,
|
15 |
+
"attention_heads": 32,
|
16 |
+
"bad_words_ids": null,
|
17 |
+
"begin_suppress_tokens": null,
|
18 |
+
"bos_token_id": 0,
|
19 |
+
"chunk_size_feed_forward": 0,
|
20 |
+
"cross_attention_hidden_size": null,
|
21 |
+
"decoder_start_token_id": null,
|
22 |
+
"diversity_penalty": 0.0,
|
23 |
+
"do_sample": false,
|
24 |
+
"dropout": 0.1,
|
25 |
+
"early_stopping": false,
|
26 |
+
"embed_dim": 2048,
|
27 |
+
"encoder_no_repeat_ngram_size": 0,
|
28 |
+
"eos_token_id": 2,
|
29 |
+
"exponential_decay_length_penalty": null,
|
30 |
+
"ffn_dim": 8192,
|
31 |
+
"finetuning_task": null,
|
32 |
+
"forced_bos_token_id": null,
|
33 |
+
"forced_eos_token_id": null,
|
34 |
+
"id2label": {
|
35 |
+
"0": "LABEL_0",
|
36 |
+
"1": "LABEL_1"
|
37 |
+
},
|
38 |
+
"init_std": 0.02,
|
39 |
+
"is_decoder": false,
|
40 |
+
"is_encoder_decoder": false,
|
41 |
+
"label2id": {
|
42 |
+
"LABEL_0": 0,
|
43 |
+
"LABEL_1": 1
|
44 |
+
},
|
45 |
+
"layer_norm_eps": 1e-05,
|
46 |
+
"layerdrop": 0.0,
|
47 |
+
"layers": 24,
|
48 |
+
"length_penalty": 1.0,
|
49 |
+
"max_length": 20,
|
50 |
+
"max_position_embeddings": 2048,
|
51 |
+
"min_length": 0,
|
52 |
+
"model_type": "kosmos_2_text_model",
|
53 |
+
"no_repeat_ngram_size": 0,
|
54 |
+
"num_beam_groups": 1,
|
55 |
+
"num_beams": 1,
|
56 |
+
"num_return_sequences": 1,
|
57 |
+
"output_attentions": false,
|
58 |
+
"output_hidden_states": false,
|
59 |
+
"output_scores": false,
|
60 |
+
"pad_token_id": 1,
|
61 |
+
"prefix": null,
|
62 |
+
"problem_type": null,
|
63 |
+
"pruned_heads": {},
|
64 |
+
"remove_invalid_values": false,
|
65 |
+
"repetition_penalty": 1.0,
|
66 |
+
"return_dict": true,
|
67 |
+
"return_dict_in_generate": false,
|
68 |
+
"scale_embedding": true,
|
69 |
+
"sep_token_id": null,
|
70 |
+
"suppress_tokens": null,
|
71 |
+
"task_specific_params": null,
|
72 |
+
"temperature": 1.0,
|
73 |
+
"tf_legacy_loss": false,
|
74 |
+
"tie_encoder_decoder": false,
|
75 |
+
"tie_word_embeddings": true,
|
76 |
+
"tokenizer_class": null,
|
77 |
+
"top_k": 50,
|
78 |
+
"top_p": 1.0,
|
79 |
+
"torch_dtype": null,
|
80 |
+
"torchscript": false,
|
81 |
+
"typical_p": 1.0,
|
82 |
+
"use_bfloat16": false,
|
83 |
+
"use_cache": true,
|
84 |
+
"vocab_size": 65037
|
85 |
+
},
|
86 |
+
"torch_dtype": "float32",
|
87 |
+
"transformers_version": "4.42.4",
|
88 |
+
"vision_config": {
|
89 |
+
"_name_or_path": "",
|
90 |
+
"add_cross_attention": false,
|
91 |
+
"architectures": null,
|
92 |
+
"attention_dropout": 0.0,
|
93 |
+
"bad_words_ids": null,
|
94 |
+
"begin_suppress_tokens": null,
|
95 |
+
"bos_token_id": null,
|
96 |
+
"chunk_size_feed_forward": 0,
|
97 |
+
"cross_attention_hidden_size": null,
|
98 |
+
"decoder_start_token_id": null,
|
99 |
+
"diversity_penalty": 0.0,
|
100 |
+
"do_sample": false,
|
101 |
+
"early_stopping": false,
|
102 |
+
"encoder_no_repeat_ngram_size": 0,
|
103 |
+
"eos_token_id": null,
|
104 |
+
"exponential_decay_length_penalty": null,
|
105 |
+
"finetuning_task": null,
|
106 |
+
"forced_bos_token_id": null,
|
107 |
+
"forced_eos_token_id": null,
|
108 |
+
"hidden_act": "quick_gelu",
|
109 |
+
"hidden_size": 1024,
|
110 |
+
"id2label": {
|
111 |
+
"0": "LABEL_0",
|
112 |
+
"1": "LABEL_1"
|
113 |
+
},
|
114 |
+
"image_size": 224,
|
115 |
+
"initializer_factor": 1.0,
|
116 |
+
"initializer_range": 0.02,
|
117 |
+
"intermediate_size": 4096,
|
118 |
+
"is_decoder": false,
|
119 |
+
"is_encoder_decoder": false,
|
120 |
+
"label2id": {
|
121 |
+
"LABEL_0": 0,
|
122 |
+
"LABEL_1": 1
|
123 |
+
},
|
124 |
+
"layer_norm_eps": 1e-05,
|
125 |
+
"length_penalty": 1.0,
|
126 |
+
"max_length": 20,
|
127 |
+
"min_length": 0,
|
128 |
+
"model_type": "kosmos_2_vision_model",
|
129 |
+
"no_repeat_ngram_size": 0,
|
130 |
+
"num_attention_heads": 16,
|
131 |
+
"num_beam_groups": 1,
|
132 |
+
"num_beams": 1,
|
133 |
+
"num_channels": 3,
|
134 |
+
"num_hidden_layers": 24,
|
135 |
+
"num_return_sequences": 1,
|
136 |
+
"output_attentions": false,
|
137 |
+
"output_hidden_states": false,
|
138 |
+
"output_scores": false,
|
139 |
+
"pad_token_id": null,
|
140 |
+
"patch_size": 14,
|
141 |
+
"prefix": null,
|
142 |
+
"problem_type": null,
|
143 |
+
"pruned_heads": {},
|
144 |
+
"remove_invalid_values": false,
|
145 |
+
"repetition_penalty": 1.0,
|
146 |
+
"return_dict": true,
|
147 |
+
"return_dict_in_generate": false,
|
148 |
+
"sep_token_id": null,
|
149 |
+
"suppress_tokens": null,
|
150 |
+
"task_specific_params": null,
|
151 |
+
"temperature": 1.0,
|
152 |
+
"tf_legacy_loss": false,
|
153 |
+
"tie_encoder_decoder": false,
|
154 |
+
"tie_word_embeddings": true,
|
155 |
+
"tokenizer_class": null,
|
156 |
+
"top_k": 50,
|
157 |
+
"top_p": 1.0,
|
158 |
+
"torch_dtype": null,
|
159 |
+
"torchscript": false,
|
160 |
+
"typical_p": 1.0,
|
161 |
+
"use_bfloat16": false
|
162 |
+
}
|
163 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 0,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"pad_token_id": 1,
|
6 |
+
"transformers_version": "4.42.4"
|
7 |
+
}
|
model-00001-of-00002.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b87b9cc0a0998c949a5cb9256a595cffd6ece661329ba42333c9091cfc3d3d7
|
3 |
+
size 4999738624
|
model-00002-of-00002.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55059ed4eddb3d49de588da7bea3f3167b46039a74ed5dcba63bda17d8fe132b
|
3 |
+
size 1658313704
|
model.safetensors.index.json
ADDED
@@ -0,0 +1,892 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 6657941504
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"image_to_text_projection.dense.bias": "model-00002-of-00002.safetensors",
|
7 |
+
"image_to_text_projection.dense.weight": "model-00002-of-00002.safetensors",
|
8 |
+
"image_to_text_projection.latent_query": "model-00002-of-00002.safetensors",
|
9 |
+
"image_to_text_projection.x_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
10 |
+
"image_to_text_projection.x_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
11 |
+
"image_to_text_projection.x_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
12 |
+
"image_to_text_projection.x_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
13 |
+
"image_to_text_projection.x_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
14 |
+
"image_to_text_projection.x_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
15 |
+
"image_to_text_projection.x_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
16 |
+
"image_to_text_projection.x_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
17 |
+
"text_model.model.embed_tokens.weight": "model-00001-of-00002.safetensors",
|
18 |
+
"text_model.model.layer_norm.bias": "model-00002-of-00002.safetensors",
|
19 |
+
"text_model.model.layer_norm.weight": "model-00002-of-00002.safetensors",
|
20 |
+
"text_model.model.layers.0.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
21 |
+
"text_model.model.layers.0.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
22 |
+
"text_model.model.layers.0.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
23 |
+
"text_model.model.layers.0.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
24 |
+
"text_model.model.layers.0.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
25 |
+
"text_model.model.layers.0.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
26 |
+
"text_model.model.layers.0.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
27 |
+
"text_model.model.layers.0.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
28 |
+
"text_model.model.layers.0.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
29 |
+
"text_model.model.layers.0.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
30 |
+
"text_model.model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
31 |
+
"text_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
32 |
+
"text_model.model.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
33 |
+
"text_model.model.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
34 |
+
"text_model.model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
35 |
+
"text_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
36 |
+
"text_model.model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
37 |
+
"text_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
38 |
+
"text_model.model.layers.0.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
39 |
+
"text_model.model.layers.0.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
40 |
+
"text_model.model.layers.1.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
41 |
+
"text_model.model.layers.1.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
42 |
+
"text_model.model.layers.1.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
43 |
+
"text_model.model.layers.1.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
44 |
+
"text_model.model.layers.1.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
45 |
+
"text_model.model.layers.1.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
46 |
+
"text_model.model.layers.1.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
47 |
+
"text_model.model.layers.1.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
48 |
+
"text_model.model.layers.1.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
49 |
+
"text_model.model.layers.1.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
50 |
+
"text_model.model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
51 |
+
"text_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
52 |
+
"text_model.model.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
53 |
+
"text_model.model.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
54 |
+
"text_model.model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
55 |
+
"text_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
56 |
+
"text_model.model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
57 |
+
"text_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
58 |
+
"text_model.model.layers.1.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
59 |
+
"text_model.model.layers.1.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
60 |
+
"text_model.model.layers.10.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
61 |
+
"text_model.model.layers.10.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
62 |
+
"text_model.model.layers.10.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
63 |
+
"text_model.model.layers.10.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
64 |
+
"text_model.model.layers.10.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
65 |
+
"text_model.model.layers.10.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
66 |
+
"text_model.model.layers.10.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
67 |
+
"text_model.model.layers.10.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
68 |
+
"text_model.model.layers.10.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
69 |
+
"text_model.model.layers.10.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
70 |
+
"text_model.model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
71 |
+
"text_model.model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
72 |
+
"text_model.model.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
73 |
+
"text_model.model.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
74 |
+
"text_model.model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
75 |
+
"text_model.model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
76 |
+
"text_model.model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
77 |
+
"text_model.model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
78 |
+
"text_model.model.layers.10.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
79 |
+
"text_model.model.layers.10.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
80 |
+
"text_model.model.layers.11.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
81 |
+
"text_model.model.layers.11.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
82 |
+
"text_model.model.layers.11.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
83 |
+
"text_model.model.layers.11.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
84 |
+
"text_model.model.layers.11.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
85 |
+
"text_model.model.layers.11.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
86 |
+
"text_model.model.layers.11.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
87 |
+
"text_model.model.layers.11.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
88 |
+
"text_model.model.layers.11.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
89 |
+
"text_model.model.layers.11.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
90 |
+
"text_model.model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
91 |
+
"text_model.model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
92 |
+
"text_model.model.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
93 |
+
"text_model.model.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
94 |
+
"text_model.model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
95 |
+
"text_model.model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
96 |
+
"text_model.model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
97 |
+
"text_model.model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
98 |
+
"text_model.model.layers.11.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
99 |
+
"text_model.model.layers.11.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
100 |
+
"text_model.model.layers.12.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
101 |
+
"text_model.model.layers.12.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
102 |
+
"text_model.model.layers.12.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
103 |
+
"text_model.model.layers.12.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
104 |
+
"text_model.model.layers.12.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
105 |
+
"text_model.model.layers.12.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
106 |
+
"text_model.model.layers.12.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
107 |
+
"text_model.model.layers.12.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
108 |
+
"text_model.model.layers.12.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
109 |
+
"text_model.model.layers.12.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
110 |
+
"text_model.model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
111 |
+
"text_model.model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
112 |
+
"text_model.model.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
113 |
+
"text_model.model.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
114 |
+
"text_model.model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
115 |
+
"text_model.model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
116 |
+
"text_model.model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
117 |
+
"text_model.model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
118 |
+
"text_model.model.layers.12.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
119 |
+
"text_model.model.layers.12.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
120 |
+
"text_model.model.layers.13.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
121 |
+
"text_model.model.layers.13.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
122 |
+
"text_model.model.layers.13.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
123 |
+
"text_model.model.layers.13.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
124 |
+
"text_model.model.layers.13.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
125 |
+
"text_model.model.layers.13.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
126 |
+
"text_model.model.layers.13.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
127 |
+
"text_model.model.layers.13.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
128 |
+
"text_model.model.layers.13.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
129 |
+
"text_model.model.layers.13.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
130 |
+
"text_model.model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
131 |
+
"text_model.model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
132 |
+
"text_model.model.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
133 |
+
"text_model.model.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
134 |
+
"text_model.model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
135 |
+
"text_model.model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
136 |
+
"text_model.model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
137 |
+
"text_model.model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
138 |
+
"text_model.model.layers.13.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
139 |
+
"text_model.model.layers.13.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
140 |
+
"text_model.model.layers.14.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
141 |
+
"text_model.model.layers.14.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
142 |
+
"text_model.model.layers.14.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
143 |
+
"text_model.model.layers.14.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
144 |
+
"text_model.model.layers.14.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
145 |
+
"text_model.model.layers.14.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
146 |
+
"text_model.model.layers.14.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
147 |
+
"text_model.model.layers.14.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
148 |
+
"text_model.model.layers.14.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
149 |
+
"text_model.model.layers.14.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
150 |
+
"text_model.model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
151 |
+
"text_model.model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
152 |
+
"text_model.model.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
153 |
+
"text_model.model.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
154 |
+
"text_model.model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
155 |
+
"text_model.model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
156 |
+
"text_model.model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
157 |
+
"text_model.model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
158 |
+
"text_model.model.layers.14.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
159 |
+
"text_model.model.layers.14.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
160 |
+
"text_model.model.layers.15.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
161 |
+
"text_model.model.layers.15.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
162 |
+
"text_model.model.layers.15.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
163 |
+
"text_model.model.layers.15.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
164 |
+
"text_model.model.layers.15.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
165 |
+
"text_model.model.layers.15.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
166 |
+
"text_model.model.layers.15.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
167 |
+
"text_model.model.layers.15.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
168 |
+
"text_model.model.layers.15.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
169 |
+
"text_model.model.layers.15.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
170 |
+
"text_model.model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
171 |
+
"text_model.model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
172 |
+
"text_model.model.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
173 |
+
"text_model.model.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
174 |
+
"text_model.model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
175 |
+
"text_model.model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
176 |
+
"text_model.model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
177 |
+
"text_model.model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
178 |
+
"text_model.model.layers.15.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
179 |
+
"text_model.model.layers.15.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
180 |
+
"text_model.model.layers.16.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
181 |
+
"text_model.model.layers.16.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
182 |
+
"text_model.model.layers.16.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
183 |
+
"text_model.model.layers.16.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
184 |
+
"text_model.model.layers.16.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
185 |
+
"text_model.model.layers.16.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
186 |
+
"text_model.model.layers.16.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
187 |
+
"text_model.model.layers.16.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
188 |
+
"text_model.model.layers.16.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
189 |
+
"text_model.model.layers.16.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
190 |
+
"text_model.model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
191 |
+
"text_model.model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
192 |
+
"text_model.model.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
193 |
+
"text_model.model.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
194 |
+
"text_model.model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
195 |
+
"text_model.model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
196 |
+
"text_model.model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
197 |
+
"text_model.model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
198 |
+
"text_model.model.layers.16.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
199 |
+
"text_model.model.layers.16.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
200 |
+
"text_model.model.layers.17.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
201 |
+
"text_model.model.layers.17.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
202 |
+
"text_model.model.layers.17.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
203 |
+
"text_model.model.layers.17.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
204 |
+
"text_model.model.layers.17.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
205 |
+
"text_model.model.layers.17.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
206 |
+
"text_model.model.layers.17.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
207 |
+
"text_model.model.layers.17.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
208 |
+
"text_model.model.layers.17.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
209 |
+
"text_model.model.layers.17.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
210 |
+
"text_model.model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
211 |
+
"text_model.model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
212 |
+
"text_model.model.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
213 |
+
"text_model.model.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
214 |
+
"text_model.model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
215 |
+
"text_model.model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
216 |
+
"text_model.model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
217 |
+
"text_model.model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
218 |
+
"text_model.model.layers.17.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
219 |
+
"text_model.model.layers.17.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
220 |
+
"text_model.model.layers.18.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
221 |
+
"text_model.model.layers.18.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
222 |
+
"text_model.model.layers.18.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
223 |
+
"text_model.model.layers.18.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
224 |
+
"text_model.model.layers.18.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
225 |
+
"text_model.model.layers.18.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
226 |
+
"text_model.model.layers.18.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
227 |
+
"text_model.model.layers.18.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
228 |
+
"text_model.model.layers.18.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
229 |
+
"text_model.model.layers.18.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
230 |
+
"text_model.model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
231 |
+
"text_model.model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
232 |
+
"text_model.model.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
233 |
+
"text_model.model.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
234 |
+
"text_model.model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
235 |
+
"text_model.model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
236 |
+
"text_model.model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
237 |
+
"text_model.model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
238 |
+
"text_model.model.layers.18.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
239 |
+
"text_model.model.layers.18.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
240 |
+
"text_model.model.layers.19.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
241 |
+
"text_model.model.layers.19.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
242 |
+
"text_model.model.layers.19.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
243 |
+
"text_model.model.layers.19.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
244 |
+
"text_model.model.layers.19.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
245 |
+
"text_model.model.layers.19.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
246 |
+
"text_model.model.layers.19.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
247 |
+
"text_model.model.layers.19.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
248 |
+
"text_model.model.layers.19.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
249 |
+
"text_model.model.layers.19.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
250 |
+
"text_model.model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
251 |
+
"text_model.model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
252 |
+
"text_model.model.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
253 |
+
"text_model.model.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
254 |
+
"text_model.model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
255 |
+
"text_model.model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
256 |
+
"text_model.model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
257 |
+
"text_model.model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
258 |
+
"text_model.model.layers.19.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
259 |
+
"text_model.model.layers.19.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
260 |
+
"text_model.model.layers.2.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
261 |
+
"text_model.model.layers.2.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
262 |
+
"text_model.model.layers.2.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
263 |
+
"text_model.model.layers.2.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
264 |
+
"text_model.model.layers.2.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
265 |
+
"text_model.model.layers.2.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
266 |
+
"text_model.model.layers.2.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
267 |
+
"text_model.model.layers.2.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
268 |
+
"text_model.model.layers.2.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
269 |
+
"text_model.model.layers.2.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
270 |
+
"text_model.model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
271 |
+
"text_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
272 |
+
"text_model.model.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
273 |
+
"text_model.model.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
274 |
+
"text_model.model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
275 |
+
"text_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
276 |
+
"text_model.model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
277 |
+
"text_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
278 |
+
"text_model.model.layers.2.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
279 |
+
"text_model.model.layers.2.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
280 |
+
"text_model.model.layers.20.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
281 |
+
"text_model.model.layers.20.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
282 |
+
"text_model.model.layers.20.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
283 |
+
"text_model.model.layers.20.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
284 |
+
"text_model.model.layers.20.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
285 |
+
"text_model.model.layers.20.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
286 |
+
"text_model.model.layers.20.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
287 |
+
"text_model.model.layers.20.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
288 |
+
"text_model.model.layers.20.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
289 |
+
"text_model.model.layers.20.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
290 |
+
"text_model.model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
291 |
+
"text_model.model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
292 |
+
"text_model.model.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
293 |
+
"text_model.model.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
294 |
+
"text_model.model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
295 |
+
"text_model.model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
296 |
+
"text_model.model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
297 |
+
"text_model.model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
298 |
+
"text_model.model.layers.20.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
299 |
+
"text_model.model.layers.20.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
300 |
+
"text_model.model.layers.21.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
301 |
+
"text_model.model.layers.21.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
302 |
+
"text_model.model.layers.21.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
303 |
+
"text_model.model.layers.21.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
304 |
+
"text_model.model.layers.21.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
305 |
+
"text_model.model.layers.21.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
306 |
+
"text_model.model.layers.21.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
307 |
+
"text_model.model.layers.21.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
308 |
+
"text_model.model.layers.21.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
309 |
+
"text_model.model.layers.21.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
310 |
+
"text_model.model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
311 |
+
"text_model.model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
312 |
+
"text_model.model.layers.21.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
313 |
+
"text_model.model.layers.21.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
314 |
+
"text_model.model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
315 |
+
"text_model.model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
316 |
+
"text_model.model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
317 |
+
"text_model.model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
318 |
+
"text_model.model.layers.21.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
319 |
+
"text_model.model.layers.21.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
320 |
+
"text_model.model.layers.22.ffn.fc1.bias": "model-00002-of-00002.safetensors",
|
321 |
+
"text_model.model.layers.22.ffn.fc1.weight": "model-00002-of-00002.safetensors",
|
322 |
+
"text_model.model.layers.22.ffn.fc2.bias": "model-00002-of-00002.safetensors",
|
323 |
+
"text_model.model.layers.22.ffn.fc2.weight": "model-00002-of-00002.safetensors",
|
324 |
+
"text_model.model.layers.22.ffn.ffn_layernorm.bias": "model-00002-of-00002.safetensors",
|
325 |
+
"text_model.model.layers.22.ffn.ffn_layernorm.weight": "model-00002-of-00002.safetensors",
|
326 |
+
"text_model.model.layers.22.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
327 |
+
"text_model.model.layers.22.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
328 |
+
"text_model.model.layers.22.self_attn.inner_attn_ln.bias": "model-00002-of-00002.safetensors",
|
329 |
+
"text_model.model.layers.22.self_attn.inner_attn_ln.weight": "model-00002-of-00002.safetensors",
|
330 |
+
"text_model.model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
331 |
+
"text_model.model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
332 |
+
"text_model.model.layers.22.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
333 |
+
"text_model.model.layers.22.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
334 |
+
"text_model.model.layers.22.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
335 |
+
"text_model.model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
336 |
+
"text_model.model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
337 |
+
"text_model.model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
338 |
+
"text_model.model.layers.22.self_attn_layer_norm.bias": "model-00002-of-00002.safetensors",
|
339 |
+
"text_model.model.layers.22.self_attn_layer_norm.weight": "model-00002-of-00002.safetensors",
|
340 |
+
"text_model.model.layers.23.ffn.fc1.bias": "model-00002-of-00002.safetensors",
|
341 |
+
"text_model.model.layers.23.ffn.fc1.weight": "model-00002-of-00002.safetensors",
|
342 |
+
"text_model.model.layers.23.ffn.fc2.bias": "model-00002-of-00002.safetensors",
|
343 |
+
"text_model.model.layers.23.ffn.fc2.weight": "model-00002-of-00002.safetensors",
|
344 |
+
"text_model.model.layers.23.ffn.ffn_layernorm.bias": "model-00002-of-00002.safetensors",
|
345 |
+
"text_model.model.layers.23.ffn.ffn_layernorm.weight": "model-00002-of-00002.safetensors",
|
346 |
+
"text_model.model.layers.23.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
347 |
+
"text_model.model.layers.23.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
348 |
+
"text_model.model.layers.23.self_attn.inner_attn_ln.bias": "model-00002-of-00002.safetensors",
|
349 |
+
"text_model.model.layers.23.self_attn.inner_attn_ln.weight": "model-00002-of-00002.safetensors",
|
350 |
+
"text_model.model.layers.23.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
351 |
+
"text_model.model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
352 |
+
"text_model.model.layers.23.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
353 |
+
"text_model.model.layers.23.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
354 |
+
"text_model.model.layers.23.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
355 |
+
"text_model.model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
356 |
+
"text_model.model.layers.23.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
357 |
+
"text_model.model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
358 |
+
"text_model.model.layers.23.self_attn_layer_norm.bias": "model-00002-of-00002.safetensors",
|
359 |
+
"text_model.model.layers.23.self_attn_layer_norm.weight": "model-00002-of-00002.safetensors",
|
360 |
+
"text_model.model.layers.3.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
361 |
+
"text_model.model.layers.3.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
362 |
+
"text_model.model.layers.3.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
363 |
+
"text_model.model.layers.3.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
364 |
+
"text_model.model.layers.3.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
365 |
+
"text_model.model.layers.3.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
366 |
+
"text_model.model.layers.3.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
367 |
+
"text_model.model.layers.3.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
368 |
+
"text_model.model.layers.3.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
369 |
+
"text_model.model.layers.3.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
370 |
+
"text_model.model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
371 |
+
"text_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
372 |
+
"text_model.model.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
373 |
+
"text_model.model.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
374 |
+
"text_model.model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
375 |
+
"text_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
376 |
+
"text_model.model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
377 |
+
"text_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
378 |
+
"text_model.model.layers.3.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
379 |
+
"text_model.model.layers.3.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
380 |
+
"text_model.model.layers.4.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
381 |
+
"text_model.model.layers.4.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
382 |
+
"text_model.model.layers.4.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
383 |
+
"text_model.model.layers.4.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
384 |
+
"text_model.model.layers.4.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
385 |
+
"text_model.model.layers.4.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
386 |
+
"text_model.model.layers.4.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
387 |
+
"text_model.model.layers.4.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
388 |
+
"text_model.model.layers.4.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
389 |
+
"text_model.model.layers.4.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
390 |
+
"text_model.model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
391 |
+
"text_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
392 |
+
"text_model.model.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
393 |
+
"text_model.model.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
394 |
+
"text_model.model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
395 |
+
"text_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
396 |
+
"text_model.model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
397 |
+
"text_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
398 |
+
"text_model.model.layers.4.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
399 |
+
"text_model.model.layers.4.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
400 |
+
"text_model.model.layers.5.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
401 |
+
"text_model.model.layers.5.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
402 |
+
"text_model.model.layers.5.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
403 |
+
"text_model.model.layers.5.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
404 |
+
"text_model.model.layers.5.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
405 |
+
"text_model.model.layers.5.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
406 |
+
"text_model.model.layers.5.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
407 |
+
"text_model.model.layers.5.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
408 |
+
"text_model.model.layers.5.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
409 |
+
"text_model.model.layers.5.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
410 |
+
"text_model.model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
411 |
+
"text_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
412 |
+
"text_model.model.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
413 |
+
"text_model.model.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
414 |
+
"text_model.model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
415 |
+
"text_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
416 |
+
"text_model.model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
417 |
+
"text_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
418 |
+
"text_model.model.layers.5.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
419 |
+
"text_model.model.layers.5.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
420 |
+
"text_model.model.layers.6.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
421 |
+
"text_model.model.layers.6.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
422 |
+
"text_model.model.layers.6.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
423 |
+
"text_model.model.layers.6.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
424 |
+
"text_model.model.layers.6.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
425 |
+
"text_model.model.layers.6.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
426 |
+
"text_model.model.layers.6.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
427 |
+
"text_model.model.layers.6.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
428 |
+
"text_model.model.layers.6.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
429 |
+
"text_model.model.layers.6.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
430 |
+
"text_model.model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
431 |
+
"text_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
432 |
+
"text_model.model.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
433 |
+
"text_model.model.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
434 |
+
"text_model.model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
435 |
+
"text_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
436 |
+
"text_model.model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
437 |
+
"text_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
438 |
+
"text_model.model.layers.6.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
439 |
+
"text_model.model.layers.6.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
440 |
+
"text_model.model.layers.7.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
441 |
+
"text_model.model.layers.7.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
442 |
+
"text_model.model.layers.7.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
443 |
+
"text_model.model.layers.7.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
444 |
+
"text_model.model.layers.7.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
445 |
+
"text_model.model.layers.7.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
446 |
+
"text_model.model.layers.7.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
447 |
+
"text_model.model.layers.7.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
448 |
+
"text_model.model.layers.7.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
449 |
+
"text_model.model.layers.7.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
450 |
+
"text_model.model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
451 |
+
"text_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
452 |
+
"text_model.model.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
453 |
+
"text_model.model.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
454 |
+
"text_model.model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
455 |
+
"text_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
456 |
+
"text_model.model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
457 |
+
"text_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
458 |
+
"text_model.model.layers.7.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
459 |
+
"text_model.model.layers.7.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
460 |
+
"text_model.model.layers.8.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
461 |
+
"text_model.model.layers.8.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
462 |
+
"text_model.model.layers.8.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
463 |
+
"text_model.model.layers.8.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
464 |
+
"text_model.model.layers.8.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
465 |
+
"text_model.model.layers.8.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
466 |
+
"text_model.model.layers.8.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
467 |
+
"text_model.model.layers.8.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
468 |
+
"text_model.model.layers.8.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
469 |
+
"text_model.model.layers.8.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
470 |
+
"text_model.model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
471 |
+
"text_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
472 |
+
"text_model.model.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
473 |
+
"text_model.model.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
474 |
+
"text_model.model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
475 |
+
"text_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
476 |
+
"text_model.model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
477 |
+
"text_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
478 |
+
"text_model.model.layers.8.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
479 |
+
"text_model.model.layers.8.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
480 |
+
"text_model.model.layers.9.ffn.fc1.bias": "model-00001-of-00002.safetensors",
|
481 |
+
"text_model.model.layers.9.ffn.fc1.weight": "model-00001-of-00002.safetensors",
|
482 |
+
"text_model.model.layers.9.ffn.fc2.bias": "model-00001-of-00002.safetensors",
|
483 |
+
"text_model.model.layers.9.ffn.fc2.weight": "model-00001-of-00002.safetensors",
|
484 |
+
"text_model.model.layers.9.ffn.ffn_layernorm.bias": "model-00001-of-00002.safetensors",
|
485 |
+
"text_model.model.layers.9.ffn.ffn_layernorm.weight": "model-00001-of-00002.safetensors",
|
486 |
+
"text_model.model.layers.9.final_layer_norm.bias": "model-00001-of-00002.safetensors",
|
487 |
+
"text_model.model.layers.9.final_layer_norm.weight": "model-00001-of-00002.safetensors",
|
488 |
+
"text_model.model.layers.9.self_attn.inner_attn_ln.bias": "model-00001-of-00002.safetensors",
|
489 |
+
"text_model.model.layers.9.self_attn.inner_attn_ln.weight": "model-00001-of-00002.safetensors",
|
490 |
+
"text_model.model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
491 |
+
"text_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
492 |
+
"text_model.model.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
493 |
+
"text_model.model.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
494 |
+
"text_model.model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
495 |
+
"text_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
496 |
+
"text_model.model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
497 |
+
"text_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
498 |
+
"text_model.model.layers.9.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
|
499 |
+
"text_model.model.layers.9.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
|
500 |
+
"vision_model.model.embeddings.class_embedding": "model-00002-of-00002.safetensors",
|
501 |
+
"vision_model.model.embeddings.patch_embedding.weight": "model-00002-of-00002.safetensors",
|
502 |
+
"vision_model.model.embeddings.position_embedding.weight": "model-00002-of-00002.safetensors",
|
503 |
+
"vision_model.model.encoder.layers.0.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
504 |
+
"vision_model.model.encoder.layers.0.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
505 |
+
"vision_model.model.encoder.layers.0.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
506 |
+
"vision_model.model.encoder.layers.0.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
507 |
+
"vision_model.model.encoder.layers.0.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
508 |
+
"vision_model.model.encoder.layers.0.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
509 |
+
"vision_model.model.encoder.layers.0.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
510 |
+
"vision_model.model.encoder.layers.0.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
511 |
+
"vision_model.model.encoder.layers.0.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
512 |
+
"vision_model.model.encoder.layers.0.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
513 |
+
"vision_model.model.encoder.layers.0.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
514 |
+
"vision_model.model.encoder.layers.0.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
515 |
+
"vision_model.model.encoder.layers.0.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
516 |
+
"vision_model.model.encoder.layers.0.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
517 |
+
"vision_model.model.encoder.layers.0.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
518 |
+
"vision_model.model.encoder.layers.0.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
519 |
+
"vision_model.model.encoder.layers.1.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
520 |
+
"vision_model.model.encoder.layers.1.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
521 |
+
"vision_model.model.encoder.layers.1.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
522 |
+
"vision_model.model.encoder.layers.1.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
523 |
+
"vision_model.model.encoder.layers.1.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
524 |
+
"vision_model.model.encoder.layers.1.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
525 |
+
"vision_model.model.encoder.layers.1.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
526 |
+
"vision_model.model.encoder.layers.1.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
527 |
+
"vision_model.model.encoder.layers.1.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
528 |
+
"vision_model.model.encoder.layers.1.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
529 |
+
"vision_model.model.encoder.layers.1.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
530 |
+
"vision_model.model.encoder.layers.1.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
531 |
+
"vision_model.model.encoder.layers.1.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
532 |
+
"vision_model.model.encoder.layers.1.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
533 |
+
"vision_model.model.encoder.layers.1.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
534 |
+
"vision_model.model.encoder.layers.1.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
535 |
+
"vision_model.model.encoder.layers.10.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
536 |
+
"vision_model.model.encoder.layers.10.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
537 |
+
"vision_model.model.encoder.layers.10.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
538 |
+
"vision_model.model.encoder.layers.10.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
539 |
+
"vision_model.model.encoder.layers.10.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
540 |
+
"vision_model.model.encoder.layers.10.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
541 |
+
"vision_model.model.encoder.layers.10.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
542 |
+
"vision_model.model.encoder.layers.10.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
543 |
+
"vision_model.model.encoder.layers.10.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
544 |
+
"vision_model.model.encoder.layers.10.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
545 |
+
"vision_model.model.encoder.layers.10.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
546 |
+
"vision_model.model.encoder.layers.10.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
547 |
+
"vision_model.model.encoder.layers.10.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
548 |
+
"vision_model.model.encoder.layers.10.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
549 |
+
"vision_model.model.encoder.layers.10.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
550 |
+
"vision_model.model.encoder.layers.10.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
551 |
+
"vision_model.model.encoder.layers.11.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
552 |
+
"vision_model.model.encoder.layers.11.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
553 |
+
"vision_model.model.encoder.layers.11.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
554 |
+
"vision_model.model.encoder.layers.11.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
555 |
+
"vision_model.model.encoder.layers.11.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
556 |
+
"vision_model.model.encoder.layers.11.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
557 |
+
"vision_model.model.encoder.layers.11.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
558 |
+
"vision_model.model.encoder.layers.11.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
559 |
+
"vision_model.model.encoder.layers.11.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
560 |
+
"vision_model.model.encoder.layers.11.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
561 |
+
"vision_model.model.encoder.layers.11.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
562 |
+
"vision_model.model.encoder.layers.11.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
563 |
+
"vision_model.model.encoder.layers.11.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
564 |
+
"vision_model.model.encoder.layers.11.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
565 |
+
"vision_model.model.encoder.layers.11.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
566 |
+
"vision_model.model.encoder.layers.11.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
567 |
+
"vision_model.model.encoder.layers.12.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
568 |
+
"vision_model.model.encoder.layers.12.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
569 |
+
"vision_model.model.encoder.layers.12.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
570 |
+
"vision_model.model.encoder.layers.12.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
571 |
+
"vision_model.model.encoder.layers.12.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
572 |
+
"vision_model.model.encoder.layers.12.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
573 |
+
"vision_model.model.encoder.layers.12.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
574 |
+
"vision_model.model.encoder.layers.12.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
575 |
+
"vision_model.model.encoder.layers.12.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
576 |
+
"vision_model.model.encoder.layers.12.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
577 |
+
"vision_model.model.encoder.layers.12.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
578 |
+
"vision_model.model.encoder.layers.12.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
579 |
+
"vision_model.model.encoder.layers.12.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
580 |
+
"vision_model.model.encoder.layers.12.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
581 |
+
"vision_model.model.encoder.layers.12.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
582 |
+
"vision_model.model.encoder.layers.12.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
583 |
+
"vision_model.model.encoder.layers.13.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
584 |
+
"vision_model.model.encoder.layers.13.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
585 |
+
"vision_model.model.encoder.layers.13.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
586 |
+
"vision_model.model.encoder.layers.13.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
587 |
+
"vision_model.model.encoder.layers.13.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
588 |
+
"vision_model.model.encoder.layers.13.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
589 |
+
"vision_model.model.encoder.layers.13.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
590 |
+
"vision_model.model.encoder.layers.13.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
591 |
+
"vision_model.model.encoder.layers.13.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
592 |
+
"vision_model.model.encoder.layers.13.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
593 |
+
"vision_model.model.encoder.layers.13.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
594 |
+
"vision_model.model.encoder.layers.13.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
595 |
+
"vision_model.model.encoder.layers.13.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
596 |
+
"vision_model.model.encoder.layers.13.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
597 |
+
"vision_model.model.encoder.layers.13.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
598 |
+
"vision_model.model.encoder.layers.13.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
599 |
+
"vision_model.model.encoder.layers.14.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
600 |
+
"vision_model.model.encoder.layers.14.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
601 |
+
"vision_model.model.encoder.layers.14.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
602 |
+
"vision_model.model.encoder.layers.14.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
603 |
+
"vision_model.model.encoder.layers.14.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
604 |
+
"vision_model.model.encoder.layers.14.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
605 |
+
"vision_model.model.encoder.layers.14.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
606 |
+
"vision_model.model.encoder.layers.14.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
607 |
+
"vision_model.model.encoder.layers.14.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
608 |
+
"vision_model.model.encoder.layers.14.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
609 |
+
"vision_model.model.encoder.layers.14.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
610 |
+
"vision_model.model.encoder.layers.14.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
611 |
+
"vision_model.model.encoder.layers.14.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
612 |
+
"vision_model.model.encoder.layers.14.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
613 |
+
"vision_model.model.encoder.layers.14.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
614 |
+
"vision_model.model.encoder.layers.14.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
615 |
+
"vision_model.model.encoder.layers.15.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
616 |
+
"vision_model.model.encoder.layers.15.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
617 |
+
"vision_model.model.encoder.layers.15.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
618 |
+
"vision_model.model.encoder.layers.15.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
619 |
+
"vision_model.model.encoder.layers.15.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
620 |
+
"vision_model.model.encoder.layers.15.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
621 |
+
"vision_model.model.encoder.layers.15.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
622 |
+
"vision_model.model.encoder.layers.15.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
623 |
+
"vision_model.model.encoder.layers.15.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
624 |
+
"vision_model.model.encoder.layers.15.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
625 |
+
"vision_model.model.encoder.layers.15.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
626 |
+
"vision_model.model.encoder.layers.15.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
627 |
+
"vision_model.model.encoder.layers.15.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
628 |
+
"vision_model.model.encoder.layers.15.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
629 |
+
"vision_model.model.encoder.layers.15.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
630 |
+
"vision_model.model.encoder.layers.15.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
631 |
+
"vision_model.model.encoder.layers.16.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
632 |
+
"vision_model.model.encoder.layers.16.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
633 |
+
"vision_model.model.encoder.layers.16.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
634 |
+
"vision_model.model.encoder.layers.16.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
635 |
+
"vision_model.model.encoder.layers.16.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
636 |
+
"vision_model.model.encoder.layers.16.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
637 |
+
"vision_model.model.encoder.layers.16.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
638 |
+
"vision_model.model.encoder.layers.16.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
639 |
+
"vision_model.model.encoder.layers.16.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
640 |
+
"vision_model.model.encoder.layers.16.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
641 |
+
"vision_model.model.encoder.layers.16.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
642 |
+
"vision_model.model.encoder.layers.16.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
643 |
+
"vision_model.model.encoder.layers.16.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
644 |
+
"vision_model.model.encoder.layers.16.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
645 |
+
"vision_model.model.encoder.layers.16.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
646 |
+
"vision_model.model.encoder.layers.16.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
647 |
+
"vision_model.model.encoder.layers.17.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
648 |
+
"vision_model.model.encoder.layers.17.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
649 |
+
"vision_model.model.encoder.layers.17.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
650 |
+
"vision_model.model.encoder.layers.17.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
651 |
+
"vision_model.model.encoder.layers.17.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
652 |
+
"vision_model.model.encoder.layers.17.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
653 |
+
"vision_model.model.encoder.layers.17.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
654 |
+
"vision_model.model.encoder.layers.17.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
655 |
+
"vision_model.model.encoder.layers.17.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
656 |
+
"vision_model.model.encoder.layers.17.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
657 |
+
"vision_model.model.encoder.layers.17.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
658 |
+
"vision_model.model.encoder.layers.17.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
659 |
+
"vision_model.model.encoder.layers.17.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
660 |
+
"vision_model.model.encoder.layers.17.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
661 |
+
"vision_model.model.encoder.layers.17.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
662 |
+
"vision_model.model.encoder.layers.17.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
663 |
+
"vision_model.model.encoder.layers.18.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
664 |
+
"vision_model.model.encoder.layers.18.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
665 |
+
"vision_model.model.encoder.layers.18.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
666 |
+
"vision_model.model.encoder.layers.18.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
667 |
+
"vision_model.model.encoder.layers.18.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
668 |
+
"vision_model.model.encoder.layers.18.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
669 |
+
"vision_model.model.encoder.layers.18.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
670 |
+
"vision_model.model.encoder.layers.18.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
671 |
+
"vision_model.model.encoder.layers.18.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
672 |
+
"vision_model.model.encoder.layers.18.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
673 |
+
"vision_model.model.encoder.layers.18.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
674 |
+
"vision_model.model.encoder.layers.18.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
675 |
+
"vision_model.model.encoder.layers.18.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
676 |
+
"vision_model.model.encoder.layers.18.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
677 |
+
"vision_model.model.encoder.layers.18.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
678 |
+
"vision_model.model.encoder.layers.18.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
679 |
+
"vision_model.model.encoder.layers.19.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
680 |
+
"vision_model.model.encoder.layers.19.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
681 |
+
"vision_model.model.encoder.layers.19.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
682 |
+
"vision_model.model.encoder.layers.19.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
683 |
+
"vision_model.model.encoder.layers.19.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
684 |
+
"vision_model.model.encoder.layers.19.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
685 |
+
"vision_model.model.encoder.layers.19.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
686 |
+
"vision_model.model.encoder.layers.19.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
687 |
+
"vision_model.model.encoder.layers.19.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
688 |
+
"vision_model.model.encoder.layers.19.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
689 |
+
"vision_model.model.encoder.layers.19.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
690 |
+
"vision_model.model.encoder.layers.19.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
691 |
+
"vision_model.model.encoder.layers.19.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
692 |
+
"vision_model.model.encoder.layers.19.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
693 |
+
"vision_model.model.encoder.layers.19.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
694 |
+
"vision_model.model.encoder.layers.19.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
695 |
+
"vision_model.model.encoder.layers.2.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
696 |
+
"vision_model.model.encoder.layers.2.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
697 |
+
"vision_model.model.encoder.layers.2.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
698 |
+
"vision_model.model.encoder.layers.2.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
699 |
+
"vision_model.model.encoder.layers.2.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
700 |
+
"vision_model.model.encoder.layers.2.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
701 |
+
"vision_model.model.encoder.layers.2.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
702 |
+
"vision_model.model.encoder.layers.2.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
703 |
+
"vision_model.model.encoder.layers.2.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
704 |
+
"vision_model.model.encoder.layers.2.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
705 |
+
"vision_model.model.encoder.layers.2.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
706 |
+
"vision_model.model.encoder.layers.2.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
707 |
+
"vision_model.model.encoder.layers.2.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
708 |
+
"vision_model.model.encoder.layers.2.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
709 |
+
"vision_model.model.encoder.layers.2.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
710 |
+
"vision_model.model.encoder.layers.2.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
711 |
+
"vision_model.model.encoder.layers.20.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
712 |
+
"vision_model.model.encoder.layers.20.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
713 |
+
"vision_model.model.encoder.layers.20.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
714 |
+
"vision_model.model.encoder.layers.20.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
715 |
+
"vision_model.model.encoder.layers.20.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
716 |
+
"vision_model.model.encoder.layers.20.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
717 |
+
"vision_model.model.encoder.layers.20.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
718 |
+
"vision_model.model.encoder.layers.20.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
719 |
+
"vision_model.model.encoder.layers.20.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
720 |
+
"vision_model.model.encoder.layers.20.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
721 |
+
"vision_model.model.encoder.layers.20.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
722 |
+
"vision_model.model.encoder.layers.20.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
723 |
+
"vision_model.model.encoder.layers.20.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
724 |
+
"vision_model.model.encoder.layers.20.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
725 |
+
"vision_model.model.encoder.layers.20.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
726 |
+
"vision_model.model.encoder.layers.20.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
727 |
+
"vision_model.model.encoder.layers.21.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
728 |
+
"vision_model.model.encoder.layers.21.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
729 |
+
"vision_model.model.encoder.layers.21.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
730 |
+
"vision_model.model.encoder.layers.21.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
731 |
+
"vision_model.model.encoder.layers.21.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
732 |
+
"vision_model.model.encoder.layers.21.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
733 |
+
"vision_model.model.encoder.layers.21.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
734 |
+
"vision_model.model.encoder.layers.21.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
735 |
+
"vision_model.model.encoder.layers.21.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
736 |
+
"vision_model.model.encoder.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
737 |
+
"vision_model.model.encoder.layers.21.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
738 |
+
"vision_model.model.encoder.layers.21.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
739 |
+
"vision_model.model.encoder.layers.21.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
740 |
+
"vision_model.model.encoder.layers.21.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
741 |
+
"vision_model.model.encoder.layers.21.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
742 |
+
"vision_model.model.encoder.layers.21.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
743 |
+
"vision_model.model.encoder.layers.22.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
744 |
+
"vision_model.model.encoder.layers.22.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
745 |
+
"vision_model.model.encoder.layers.22.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
746 |
+
"vision_model.model.encoder.layers.22.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
747 |
+
"vision_model.model.encoder.layers.22.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
748 |
+
"vision_model.model.encoder.layers.22.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
749 |
+
"vision_model.model.encoder.layers.22.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
750 |
+
"vision_model.model.encoder.layers.22.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
751 |
+
"vision_model.model.encoder.layers.22.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
752 |
+
"vision_model.model.encoder.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
753 |
+
"vision_model.model.encoder.layers.22.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
754 |
+
"vision_model.model.encoder.layers.22.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
755 |
+
"vision_model.model.encoder.layers.22.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
756 |
+
"vision_model.model.encoder.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
757 |
+
"vision_model.model.encoder.layers.22.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
758 |
+
"vision_model.model.encoder.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
759 |
+
"vision_model.model.encoder.layers.23.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
760 |
+
"vision_model.model.encoder.layers.23.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
761 |
+
"vision_model.model.encoder.layers.23.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
762 |
+
"vision_model.model.encoder.layers.23.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
763 |
+
"vision_model.model.encoder.layers.23.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
764 |
+
"vision_model.model.encoder.layers.23.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
765 |
+
"vision_model.model.encoder.layers.23.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
766 |
+
"vision_model.model.encoder.layers.23.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
767 |
+
"vision_model.model.encoder.layers.23.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
768 |
+
"vision_model.model.encoder.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
769 |
+
"vision_model.model.encoder.layers.23.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
770 |
+
"vision_model.model.encoder.layers.23.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
771 |
+
"vision_model.model.encoder.layers.23.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
772 |
+
"vision_model.model.encoder.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
773 |
+
"vision_model.model.encoder.layers.23.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
774 |
+
"vision_model.model.encoder.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
775 |
+
"vision_model.model.encoder.layers.3.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
776 |
+
"vision_model.model.encoder.layers.3.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
777 |
+
"vision_model.model.encoder.layers.3.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
778 |
+
"vision_model.model.encoder.layers.3.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
779 |
+
"vision_model.model.encoder.layers.3.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
780 |
+
"vision_model.model.encoder.layers.3.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
781 |
+
"vision_model.model.encoder.layers.3.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
782 |
+
"vision_model.model.encoder.layers.3.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
783 |
+
"vision_model.model.encoder.layers.3.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
784 |
+
"vision_model.model.encoder.layers.3.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
785 |
+
"vision_model.model.encoder.layers.3.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
786 |
+
"vision_model.model.encoder.layers.3.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
787 |
+
"vision_model.model.encoder.layers.3.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
788 |
+
"vision_model.model.encoder.layers.3.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
789 |
+
"vision_model.model.encoder.layers.3.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
790 |
+
"vision_model.model.encoder.layers.3.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
791 |
+
"vision_model.model.encoder.layers.4.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
792 |
+
"vision_model.model.encoder.layers.4.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
793 |
+
"vision_model.model.encoder.layers.4.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
794 |
+
"vision_model.model.encoder.layers.4.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
795 |
+
"vision_model.model.encoder.layers.4.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
796 |
+
"vision_model.model.encoder.layers.4.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
797 |
+
"vision_model.model.encoder.layers.4.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
798 |
+
"vision_model.model.encoder.layers.4.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
799 |
+
"vision_model.model.encoder.layers.4.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
800 |
+
"vision_model.model.encoder.layers.4.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
801 |
+
"vision_model.model.encoder.layers.4.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
802 |
+
"vision_model.model.encoder.layers.4.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
803 |
+
"vision_model.model.encoder.layers.4.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
804 |
+
"vision_model.model.encoder.layers.4.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
805 |
+
"vision_model.model.encoder.layers.4.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
806 |
+
"vision_model.model.encoder.layers.4.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
807 |
+
"vision_model.model.encoder.layers.5.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
808 |
+
"vision_model.model.encoder.layers.5.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
809 |
+
"vision_model.model.encoder.layers.5.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
810 |
+
"vision_model.model.encoder.layers.5.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
811 |
+
"vision_model.model.encoder.layers.5.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
812 |
+
"vision_model.model.encoder.layers.5.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
813 |
+
"vision_model.model.encoder.layers.5.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
814 |
+
"vision_model.model.encoder.layers.5.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
815 |
+
"vision_model.model.encoder.layers.5.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
816 |
+
"vision_model.model.encoder.layers.5.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
817 |
+
"vision_model.model.encoder.layers.5.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
818 |
+
"vision_model.model.encoder.layers.5.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
819 |
+
"vision_model.model.encoder.layers.5.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
820 |
+
"vision_model.model.encoder.layers.5.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
821 |
+
"vision_model.model.encoder.layers.5.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
822 |
+
"vision_model.model.encoder.layers.5.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
823 |
+
"vision_model.model.encoder.layers.6.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
824 |
+
"vision_model.model.encoder.layers.6.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
825 |
+
"vision_model.model.encoder.layers.6.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
826 |
+
"vision_model.model.encoder.layers.6.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
827 |
+
"vision_model.model.encoder.layers.6.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
828 |
+
"vision_model.model.encoder.layers.6.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
829 |
+
"vision_model.model.encoder.layers.6.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
830 |
+
"vision_model.model.encoder.layers.6.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
831 |
+
"vision_model.model.encoder.layers.6.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
832 |
+
"vision_model.model.encoder.layers.6.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
833 |
+
"vision_model.model.encoder.layers.6.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
834 |
+
"vision_model.model.encoder.layers.6.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
835 |
+
"vision_model.model.encoder.layers.6.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
836 |
+
"vision_model.model.encoder.layers.6.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
837 |
+
"vision_model.model.encoder.layers.6.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
838 |
+
"vision_model.model.encoder.layers.6.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
839 |
+
"vision_model.model.encoder.layers.7.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
840 |
+
"vision_model.model.encoder.layers.7.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
841 |
+
"vision_model.model.encoder.layers.7.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
842 |
+
"vision_model.model.encoder.layers.7.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
843 |
+
"vision_model.model.encoder.layers.7.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
844 |
+
"vision_model.model.encoder.layers.7.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
845 |
+
"vision_model.model.encoder.layers.7.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
846 |
+
"vision_model.model.encoder.layers.7.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
847 |
+
"vision_model.model.encoder.layers.7.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
848 |
+
"vision_model.model.encoder.layers.7.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
849 |
+
"vision_model.model.encoder.layers.7.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
850 |
+
"vision_model.model.encoder.layers.7.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
851 |
+
"vision_model.model.encoder.layers.7.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
852 |
+
"vision_model.model.encoder.layers.7.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
853 |
+
"vision_model.model.encoder.layers.7.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
854 |
+
"vision_model.model.encoder.layers.7.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
855 |
+
"vision_model.model.encoder.layers.8.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
856 |
+
"vision_model.model.encoder.layers.8.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
857 |
+
"vision_model.model.encoder.layers.8.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
858 |
+
"vision_model.model.encoder.layers.8.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
859 |
+
"vision_model.model.encoder.layers.8.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
860 |
+
"vision_model.model.encoder.layers.8.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
861 |
+
"vision_model.model.encoder.layers.8.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
862 |
+
"vision_model.model.encoder.layers.8.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
863 |
+
"vision_model.model.encoder.layers.8.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
864 |
+
"vision_model.model.encoder.layers.8.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
865 |
+
"vision_model.model.encoder.layers.8.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
866 |
+
"vision_model.model.encoder.layers.8.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
867 |
+
"vision_model.model.encoder.layers.8.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
868 |
+
"vision_model.model.encoder.layers.8.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
869 |
+
"vision_model.model.encoder.layers.8.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
870 |
+
"vision_model.model.encoder.layers.8.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
871 |
+
"vision_model.model.encoder.layers.9.layer_norm1.bias": "model-00002-of-00002.safetensors",
|
872 |
+
"vision_model.model.encoder.layers.9.layer_norm1.weight": "model-00002-of-00002.safetensors",
|
873 |
+
"vision_model.model.encoder.layers.9.layer_norm2.bias": "model-00002-of-00002.safetensors",
|
874 |
+
"vision_model.model.encoder.layers.9.layer_norm2.weight": "model-00002-of-00002.safetensors",
|
875 |
+
"vision_model.model.encoder.layers.9.mlp.fc1.bias": "model-00002-of-00002.safetensors",
|
876 |
+
"vision_model.model.encoder.layers.9.mlp.fc1.weight": "model-00002-of-00002.safetensors",
|
877 |
+
"vision_model.model.encoder.layers.9.mlp.fc2.bias": "model-00002-of-00002.safetensors",
|
878 |
+
"vision_model.model.encoder.layers.9.mlp.fc2.weight": "model-00002-of-00002.safetensors",
|
879 |
+
"vision_model.model.encoder.layers.9.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
|
880 |
+
"vision_model.model.encoder.layers.9.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
|
881 |
+
"vision_model.model.encoder.layers.9.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
882 |
+
"vision_model.model.encoder.layers.9.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
883 |
+
"vision_model.model.encoder.layers.9.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
|
884 |
+
"vision_model.model.encoder.layers.9.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
885 |
+
"vision_model.model.encoder.layers.9.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
|
886 |
+
"vision_model.model.encoder.layers.9.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
|
887 |
+
"vision_model.model.post_layernorm.bias": "model-00002-of-00002.safetensors",
|
888 |
+
"vision_model.model.post_layernorm.weight": "model-00002-of-00002.safetensors",
|
889 |
+
"vision_model.model.pre_layrnorm.bias": "model-00002-of-00002.safetensors",
|
890 |
+
"vision_model.model.pre_layrnorm.weight": "model-00002-of-00002.safetensors"
|
891 |
+
}
|
892 |
+
}
|
preprocessor_config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"crop_size": {
|
3 |
+
"height": 224,
|
4 |
+
"width": 224
|
5 |
+
},
|
6 |
+
"do_center_crop": true,
|
7 |
+
"do_convert_rgb": true,
|
8 |
+
"do_normalize": true,
|
9 |
+
"do_rescale": true,
|
10 |
+
"do_resize": true,
|
11 |
+
"image_mean": [
|
12 |
+
0.48145466,
|
13 |
+
0.4578275,
|
14 |
+
0.40821073
|
15 |
+
],
|
16 |
+
"image_processor_type": "CLIPImageProcessor",
|
17 |
+
"image_std": [
|
18 |
+
0.26862954,
|
19 |
+
0.26130258,
|
20 |
+
0.27577711
|
21 |
+
],
|
22 |
+
"processor_class": "Kosmos2Processor",
|
23 |
+
"resample": 3,
|
24 |
+
"rescale_factor": 0.00392156862745098,
|
25 |
+
"size": {
|
26 |
+
"height": 224,
|
27 |
+
"width": 224
|
28 |
+
}
|
29 |
+
}
|
processor_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_patch_index_tokens": 1024,
|
3 |
+
"processor_class": "Kosmos2Processor"
|
4 |
+
}
|
sentencepiece.bpe.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a60b4d1d1d8f70c8b2569c94540d4d9b7c694fd32e7a428ad0dcffaafaa3beb
|
3 |
+
size 1363614
|
special_tokens_map.json
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<s>",
|
4 |
+
"<pad>",
|
5 |
+
"</s>",
|
6 |
+
"<unk>",
|
7 |
+
"<mask>"
|
8 |
+
],
|
9 |
+
"bos_token": {
|
10 |
+
"content": "<s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"cls_token": {
|
17 |
+
"content": "<s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"eos_token": {
|
24 |
+
"content": "</s>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"mask_token": {
|
31 |
+
"content": "<mask>",
|
32 |
+
"lstrip": true,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
},
|
37 |
+
"pad_token": {
|
38 |
+
"content": "<pad>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false
|
43 |
+
},
|
44 |
+
"sep_token": {
|
45 |
+
"content": "</s>",
|
46 |
+
"lstrip": false,
|
47 |
+
"normalized": false,
|
48 |
+
"rstrip": false,
|
49 |
+
"single_word": false
|
50 |
+
},
|
51 |
+
"unk_token": {
|
52 |
+
"content": "<unk>",
|
53 |
+
"lstrip": false,
|
54 |
+
"normalized": false,
|
55 |
+
"rstrip": false,
|
56 |
+
"single_word": false
|
57 |
+
}
|
58 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96eb1da4bccf34c021e586af8926c3717978f94b043f4411bd4a731304793434
|
3 |
+
size 5048
|