Update README.md
Browse files
README.md
CHANGED
@@ -105,4 +105,8 @@ dpo_trainer = DPOTrainer(
|
|
105 |
beta=0.1,
|
106 |
max_prompt_length=1024,
|
107 |
max_length=1536,
|
108 |
-
)
|
|
|
|
|
|
|
|
|
|
105 |
beta=0.1,
|
106 |
max_prompt_length=1024,
|
107 |
max_length=1536,
|
108 |
+
)
|
109 |
+
|
110 |
+
|
111 |
+
EDIT: Still waiting for the Open-LLM benchmark results to come back in, but...:
|
112 |
+
According to the few tests I ran on it myself the new "CultriX/MistralTrix-SLERP" should beat this model at only 7.42B!
|