Update config.json
Browse files- config.json +3 -3
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"CXRLLAVAModel"
|
5 |
],
|
@@ -25,7 +25,7 @@
|
|
25 |
"std": 0.3821719215686275
|
26 |
},
|
27 |
"llama": {
|
28 |
-
"_name_or_path": "
|
29 |
"add_cross_attention": false,
|
30 |
"architectures": [
|
31 |
"LlamaForCausalLM"
|
@@ -104,7 +104,7 @@
|
|
104 |
"vocab_size": 32000
|
105 |
},
|
106 |
"llama_model_dtype": "bf16",
|
107 |
-
"llama_model_path": "
|
108 |
"mm_projector_dim": 1024,
|
109 |
"mm_projector_dtype": "fp32",
|
110 |
"mm_projector_path": null,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "CXR-LLAVA-v2",
|
3 |
"architectures": [
|
4 |
"CXRLLAVAModel"
|
5 |
],
|
|
|
25 |
"std": 0.3821719215686275
|
26 |
},
|
27 |
"llama": {
|
28 |
+
"_name_or_path": "CXR-LLAVA-v2",
|
29 |
"add_cross_attention": false,
|
30 |
"architectures": [
|
31 |
"LlamaForCausalLM"
|
|
|
104 |
"vocab_size": 32000
|
105 |
},
|
106 |
"llama_model_dtype": "bf16",
|
107 |
+
"llama_model_path": "CXR-LLAVA-v2",
|
108 |
"mm_projector_dim": 1024,
|
109 |
"mm_projector_dtype": "fp32",
|
110 |
"mm_projector_path": null,
|