GunaKoppula
commited on
Commit
•
95f1091
1
Parent(s):
deeed66
Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/p
|
|
17 |
2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
|
18 |
4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
|
19 |
5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
|
20 |
-
6. Finetuned Model: [
|
21 |
|
22 |
|
23 |
### Model Sources
|
|
|
17 |
2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
|
18 |
4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
|
19 |
5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
|
20 |
+
6. Finetuned Model: [GunaKoppula/Llava-Phi2](https://huggingface.co/GunaKoppula/Llava-Phi2)
|
21 |
|
22 |
|
23 |
### Model Sources
|