Max

moock
Β·

AI & ML interests

None yet

Recent Activity

liked a model about 1 month ago
ostris/OpenFLUX.1
liked a Space about 2 months ago
philschmid/llm-pricing
View all activity

Organizations

moock's activity

liked a Space about 2 months ago
Reacted to clem's post with πŸš€ 5 months ago
view post
Post
5770
5,000 new repos (models, datasets, spaces) are created EVERY DAY on HF now. The community is amazing!
replied to lunarflu's post 6 months ago
view reply

It would be fun to have a prediction of my future daily activities πŸͺ„

Reacted to lunarflu's post with πŸ”₯ 6 months ago
view post
Post
1902
cooking up something....anyone interested in a daily activity tracker for HF?
Β·
Reacted to singhsidhukuldeep's post with πŸ‘ 6 months ago
view post
Post
2085
🎭 You picked an LLM for your work but then you find out it hallucinates! πŸ€–

πŸ€” Your first thought might be to fine-tune it on more training data.... but should you? πŸ› οΈ

πŸ“œ This is what @Google is exploring in the paper "Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?" πŸ•΅οΈβ€β™‚οΈ

πŸ“˜ When LLMs undergo supervised fine-tuning with new factual knowledge not present in their initial training data, there is a risk they might "hallucinate" or produce factually incorrect information. 🚨

πŸ” The paper investigates how fine-tuning LLMs with new facts influences their ability to leverage pre-existing knowledge and the extent to which they generate errors. πŸ“Š

βš™οΈTechnical Setup:

πŸ”§ Approach: They introduce a system named SliCK (this stands for Sampling-based Categorization of Knowledge, don't even bother understanding how) to categorize knowledge into four levels (HighlyKnown, MaybeKnown, WeaklyKnown, and Unknown) based on how well the model's generated responses agree with known facts. πŸ—‚οΈ

πŸ“ Experimental Setup: The study uses a controlled setup focusing on closed-book QA, adjusting the proportion of fine-tuning examples that introduce new facts versus those that do not. πŸ§ͺ

πŸ‘‰ Here is the gist of the findings:

🚸 LLMs struggle to integrate new factual knowledge during fine-tuning, and such examples are learned slower than those consistent with the model's pre-existing knowledge. 🐒

πŸ“ˆ As LLMs learn from examples containing new knowledge, their propensity to hallucinate increases. πŸ‘»

⏱️ Early stopping during training can mitigate the risks of hallucinations by minimizing exposure to unlearned new facts. πŸ›‘

🧠 Training LLMs mostly with known examples leads to better utilization of pre-existing knowledge, whereas examples introducing new knowledge increase the risk of generating incorrect information. πŸ—οΈ

πŸ“„ Paper: Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? (2405.05904) πŸ“š
  • 2 replies
Β·
New activity in yanze/PuLID 6 months ago
liked a Space 7 months ago