InferenceIllusionist
commited on
Commit
•
a0ebbc7
1
Parent(s):
31b4e1e
Update README.md
Browse files
README.md
CHANGED
@@ -15,10 +15,13 @@ datasets:
|
|
15 |
<img src="https://i.imgur.com/pbPbqq0.jpeg" width="550"/>
|
16 |
|
17 |
An initial foray into the world of fine-tuning. The goal of this release was to amplify the quality of this model's responses, especially when used in vision use cases*
|
18 |
-
*(Requires [mistral-7b-mmproj-v1.5-Q4_1](https://huggingface.co/koboldcpp/mmproj/resolve/main/mistral-7b-mmproj-v1.5-Q4_1.gguf?download=true) file in Kobold)
|
19 |
|
20 |
|
21 |
### Notes & Methodology
|
22 |
* [Excalibur-7b](https://huggingface.co/InferenceIllusionist/Excalibur-7b) fine-tuned with Direct Preference Optimization (DPO) using Intel/orca_dpo_pairs
|
23 |
* This is a quick experiment to determine the impact of DPO finetuning on the original base model
|
24 |
-
* Executed for a little over an hour on a single A100
|
|
|
|
|
|
|
|
|
|
15 |
<img src="https://i.imgur.com/pbPbqq0.jpeg" width="550"/>
|
16 |
|
17 |
An initial foray into the world of fine-tuning. The goal of this release was to amplify the quality of this model's responses, especially when used in vision use cases*
|
|
|
18 |
|
19 |
|
20 |
### Notes & Methodology
|
21 |
* [Excalibur-7b](https://huggingface.co/InferenceIllusionist/Excalibur-7b) fine-tuned with Direct Preference Optimization (DPO) using Intel/orca_dpo_pairs
|
22 |
* This is a quick experiment to determine the impact of DPO finetuning on the original base model
|
23 |
+
* Executed for a little over an hour on a single A100
|
24 |
+
|
25 |
+
|
26 |
+
|
27 |
+
*Requires [mistral-7b-mmproj-v1.5-Q4_1](https://huggingface.co/koboldcpp/mmproj/resolve/main/mistral-7b-mmproj-v1.5-Q4_1.gguf?download=true) file to be loaded in Kobold
|