๐ Super @DeepMount00 just released ๐๐ฒ๐บ๐บ๐ฎ_๐ค๐_๐๐ง๐_๐๐ฏ ๐น๐ฒ๐ฎ๐ฑ๐ถ๐ป๐ด the ๐ฅ๐๐ ๐๐ฎ๐๐ธ on the Italian ๐๐๐ _๐๐ง๐_๐๐๐๐๐๐ฅ๐๐ข๐๐ฅ๐. The model is a fine tuned version of Gemma 2B. Model details: DeepMount00/Gemma_QA_ITA_v3 Explore the full RAG section rankings here: FinancialSupport/open_ita_llm_leaderboard on section Classifica RAG
On evaluating fine tuned 7B Italian open source LLMs I have collected many data points and I created a super simple explorative analyses. My hypothesis based on data are:
- mmlu is hard to improve when fine tuning a base model on a different language - fine tuning also on single GPUs can improve by 5% to 10% the base model on common tasks but a lot more on specific cases with the right training time and data - fine tuning can specialize well but at cost of loosing some foundational knowledge.