Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
singhsidhukuldeepΒ 
posted an update 15 days ago
Post
2063
🎭 You picked an LLM for your work but then you find out it hallucinates! πŸ€–

πŸ€” Your first thought might be to fine-tune it on more training data.... but should you? πŸ› οΈ

πŸ“œ This is what @Google is exploring in the paper "Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?" πŸ•΅οΈβ€β™‚οΈ

πŸ“˜ When LLMs undergo supervised fine-tuning with new factual knowledge not present in their initial training data, there is a risk they might "hallucinate" or produce factually incorrect information. 🚨

πŸ” The paper investigates how fine-tuning LLMs with new facts influences their ability to leverage pre-existing knowledge and the extent to which they generate errors. πŸ“Š

βš™οΈTechnical Setup:

πŸ”§ Approach: They introduce a system named SliCK (this stands for Sampling-based Categorization of Knowledge, don't even bother understanding how) to categorize knowledge into four levels (HighlyKnown, MaybeKnown, WeaklyKnown, and Unknown) based on how well the model's generated responses agree with known facts. πŸ—‚οΈ

πŸ“ Experimental Setup: The study uses a controlled setup focusing on closed-book QA, adjusting the proportion of fine-tuning examples that introduce new facts versus those that do not. πŸ§ͺ

πŸ‘‰ Here is the gist of the findings:

🚸 LLMs struggle to integrate new factual knowledge during fine-tuning, and such examples are learned slower than those consistent with the model's pre-existing knowledge. 🐒

πŸ“ˆ As LLMs learn from examples containing new knowledge, their propensity to hallucinate increases. πŸ‘»

⏱️ Early stopping during training can mitigate the risks of hallucinations by minimizing exposure to unlearned new facts. πŸ›‘

🧠 Training LLMs mostly with known examples leads to better utilization of pre-existing knowledge, whereas examples introducing new knowledge increase the risk of generating incorrect information. πŸ—οΈ

πŸ“„ Paper: Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? (2405.05904) πŸ“š

Hi kuldeep where are you now live?

I read this 2 years ago, and more than 10 other papers saying this in different ways.