Post
1631
Please... feed this Llama some Sauerkraut! ๐ฒ
Said and done. Here it is. Our Sauerkraut Version of the strong Llama3-8b by Meta. Released from HANNOVER MESSE, just in front of meta booth.
VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
According to benchmarks (LM-Evaluation-Harness 0.4.2), our #SauerkrautLM Dataset and fine-tuning pipeline improved the Model noticeably (AVG = 74,57), especially Reasoning and Common Sense capabilities.
Again we provide some more detail on the whole process:
โ Original model: Llama-3-8b-Instruct
โ Training Duration: 12 hours
โ Training procedure: 2-staged DPO
โ Trained data: 70k (first stage) and 20k (second stage)
โ GPU: 4x RTX6000 ADA
โ New model: Llama-3-SauerkrautLM-8b-Instruct
โ Total training costs: 54,72 Dollar ๐ด - RunPod FTW (excluding synthesizing data, curating data, benchmarks, error handling, testing)
See our model card on Hugging Face for more details: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
There will be more details on benchmarks during the next days.
Said and done. Here it is. Our Sauerkraut Version of the strong Llama3-8b by Meta. Released from HANNOVER MESSE, just in front of meta booth.
VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
According to benchmarks (LM-Evaluation-Harness 0.4.2), our #SauerkrautLM Dataset and fine-tuning pipeline improved the Model noticeably (AVG = 74,57), especially Reasoning and Common Sense capabilities.
Again we provide some more detail on the whole process:
โ Original model: Llama-3-8b-Instruct
โ Training Duration: 12 hours
โ Training procedure: 2-staged DPO
โ Trained data: 70k (first stage) and 20k (second stage)
โ GPU: 4x RTX6000 ADA
โ New model: Llama-3-SauerkrautLM-8b-Instruct
โ Total training costs: 54,72 Dollar ๐ด - RunPod FTW (excluding synthesizing data, curating data, benchmarks, error handling, testing)
See our model card on Hugging Face for more details: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
There will be more details on benchmarks during the next days.