--- license: mit language: - de - en --- ![SauerkrautLM-Phi-3-medium](https://vago-solutions.ai/wp-content/uploads/2024/06/SauerkrautLM-phi3-medium.png "SauerkrautLM-Phi-3-medium") ## VAGO solutions SauerkrautLM-Phi-3-medium Introducing **SauerkrautLM-Phi-3-medium** – our Sauerkraut version of the powerful [unsloth/Phi-3-medium-4k-instruct](https://huggingface.co/unsloth/Phi-3-medium-4k-instruct)! - Aligned with DPO using [**Spectrum**](https://github.com/cognitivecomputations/spectrum) QLoRA (by Eric Hartford, Lucas Atkins, Fernando Fernandes Neto and David Golchinfar) **targeting 50% of the layers.** # Table of Contents 1. [Overview of all SauerkrautLM-Phi-3-medium](#all-SauerkrautLM-Phi-3-medium) 2. [Model Details](#model-details) - [Training procedure](#training-procedure) 3. [Evaluation](#evaluation) 5. [Disclaimer](#disclaimer) 6. [Contact](#contact) 7. [Collaborations](#collaborations) 8. [Acknowledgement](#acknowledgement) ## All SauerkrautLM-Phi-3-medium | Model | HF | EXL2 | GGUF | AWQ | |-------|-------|-------|-------|-------| | SauerkrautLM-Phi-3-medium | [Link](https://huggingface.co/VAGOsolutions/SauerkrautLM-Phi-3-medium) | coming soon | coming soon | coming soon | ## Model Details **SauerkrautLM-Phi-3-medium** - **Model Type:** SauerkrautLM-Phi-3-medium is a finetuned Model based on [unsloth/Phi-3-medium-4k-instruct](https://huggingface.co/unsloth/Phi-3-medium-4k-instruct) - **Language(s):** German, English - **License:** MIT - **Contact:** [VAGO solutions](https://vago-solutions.ai) ### Training procedure: - We trained this model with [**Spectrum**](https://github.com/cognitivecomputations/spectrum) QLoRA DPO Fine-Tuning for 1 epoch with 70k samples targeting 50% of the layers with a high Learningrate of 5e-04. This relatively high learning rate was feasible due to the selective targeting of layers; had we applied this rate to all layers, the gradients would have exploded. **Fine-Tuning Details** Epochs: 1 Data Size: 70,000 samples Targeted Layers: 50% Learning Rate: 5e-04 Warm-up Ratio: 0.03 The strategy of targeting only half of the layers also enabled us to use a very low warm-up ratio of 0.03, contributing to the overall stability of the fine-tuning process. **Results** This fine-tuning approach resulted in a noticeable improvement in the model's reasoning capabilities. The model's performance was evaluated using a variety of benchmark suites, including the newly introduced [MixEval](https://mixeval.github.io/), which shows a 96% correlation with Chatbot Arena. MixEval uses regular updated test data, providing a reliable benchmark for model performance. ## Evaluation **Open LLM Leaderboard and German RAG:** ![SauerkrautLM-Phi-3-medium_h6_ger_rag](https://vago-solutions.ai/wp-content/uploads/2024/06/HF6-RAG.png "SauerkrautLM-Phi-3-medium_h6_ger_rag") **Mix Eval Hard** ![SauerkrautLM-Phi-3-medium_mixeval_hard](https://vago-solutions.ai/wp-content/uploads/2024/06/MixedEval.png "SauerkrautLM-Phi-3-medium_mixeval_hard") **GPT4ALL** ![SauerkrautLM-Phi-3-medium_gpt4all](https://vago-solutions.ai/wp-content/uploads/2024/06/GPT4ALL.png "SauerkrautLM-Phi-3-medium_gpt4all") **AGIEval** ![SauerkrautLM-Phi-3-medium_agieval](https://vago-solutions.ai/wp-content/uploads/2024/06/AgiEval.png "SauerkrautLM-Phi-3-medium_agieval") ## Disclaimer We must inform users that despite our best efforts in data cleansing, the possibility of uncensored content slipping through cannot be entirely ruled out. However, we cannot guarantee consistently appropriate behavior. Therefore, if you encounter any issues or come across inappropriate content, we kindly request that you inform us through the contact information provided. Additionally, it is essential to understand that the licensing of these models does not constitute legal advice. We are not held responsible for the actions of third parties who utilize our models. ## Contact If you are interested in customized LLMs for business applications, please get in contact with us via our websites. We are also grateful for your feedback and suggestions. ## Collaborations We are also keenly seeking support and investment for our startup, VAGO solutions where we continuously advance the development of robust language models designed to address a diverse range of purposes and requirements. If the prospect of collaboratively navigating future challenges excites you, we warmly invite you to reach out to us at [VAGO solutions](https://vago-solutions.ai/#Kontakt) ## Acknowledgement Many thanks to [unsloth](https://huggingface.co/unsloth/) and [Microsoft](https://huggingface.co/microsoft) for providing such valuable model to the Open-Source community.