Edit model card

SauerkrautLM-Phi-3-medium

VAGO solutions SauerkrautLM-Phi-3-medium

Introducing SauerkrautLM-Phi-3-medium – our Sauerkraut version of the powerful unsloth/Phi-3-medium-4k-instruct!

  • Aligned with DPO using Spectrum QLoRA (by Eric Hartford, Lucas Atkins, Fernando Fernandes Neto and David Golchinfar) targeting 50% of the layers.

Table of Contents

  1. Overview of all SauerkrautLM-Phi-3-medium
  2. Model Details
  3. Evaluation
  4. Disclaimer
  5. Contact
  6. Collaborations
  7. Acknowledgement

All SauerkrautLM-Phi-3-medium

Model HF EXL2 GGUF AWQ
SauerkrautLM-Phi-3-medium Link coming soon coming soon coming soon

Model Details

SauerkrautLM-Phi-3-medium

Training procedure:

  • We trained this model with Spectrum QLoRA DPO Fine-Tuning for 1 epoch with 70k samples targeting 50% of the layers with a high Learningrate of 5e-04. This relatively high learning rate was feasible due to the selective targeting of layers; had we applied this rate to all layers, the gradients would have exploded.

Fine-Tuning Details

Epochs: 1 Data Size: 70,000 samples Targeted Layers: 50% Learning Rate: 5e-04 Warm-up Ratio: 0.03 The strategy of targeting only half of the layers also enabled us to use a very low warm-up ratio of 0.03, contributing to the overall stability of the fine-tuning process.

Results

This fine-tuning approach resulted in a noticeable improvement in the model's reasoning capabilities. The model's performance was evaluated using a variety of benchmark suites, including the newly introduced MixEval, which shows a 96% correlation with Chatbot Arena. MixEval uses regular updated test data, providing a reliable benchmark for model performance.

Evaluation

Open LLM Leaderboard and German RAG:

SauerkrautLM-Phi-3-medium_h6_ger_rag

Mix Eval Hard SauerkrautLM-Phi-3-medium_mixeval_hard

GPT4ALL SauerkrautLM-Phi-3-medium_gpt4all

AGIEval

SauerkrautLM-Phi-3-medium_agieval

Disclaimer

We must inform users that despite our best efforts in data cleansing, the possibility of uncensored content slipping through cannot be entirely ruled out. However, we cannot guarantee consistently appropriate behavior. Therefore, if you encounter any issues or come across inappropriate content, we kindly request that you inform us through the contact information provided. Additionally, it is essential to understand that the licensing of these models does not constitute legal advice. We are not held responsible for the actions of third parties who utilize our models.

Contact

If you are interested in customized LLMs for business applications, please get in contact with us via our websites. We are also grateful for your feedback and suggestions.

Collaborations

We are also keenly seeking support and investment for our startup, VAGO solutions where we continuously advance the development of robust language models designed to address a diverse range of purposes and requirements. If the prospect of collaboratively navigating future challenges excites you, we warmly invite you to reach out to us at VAGO solutions

Acknowledgement

Many thanks to unsloth and Microsoft for providing such valuable model to the Open-Source community.

Downloads last month
46
Safetensors
Model size
14B params
Tensor type
BF16
·