qubite

Xdotnet

AI & ML interests

Aspiring Civil Engineer | Passionate About Sustainable Infrastructure | Problem Solver| Tech enthusiasts

Recent Activity

Reacted to m-ric's post with πŸš€ about 2 months ago
𝗔𝗱𝗱 π˜€π—Όπ˜‚π—Ώπ—°π—² π—΅π—Άπ—΄π—΅π—Ήπ—Άπ—΄π—΅π˜π—Άπ—»π—΄ π˜π—Ό π˜†π—Όπ˜‚π—Ώ π—₯π—”π—š π˜€π˜†π˜€π˜π—²π—Ί! πŸ“„πŸ’‘ RAG systems are supposed to make your LLM's answer more trustworthy, by inserting in the prompt some supporting documents from a knowledge base : we say that we're "adding some context". πŸ‘Ž But if you don't know which part of the answer has been generated based on which input tokens, it's hard to tell wether it was effectively grounded in the context knowledge or not! πŸ€” I've been working on the question: is it possible to add notes to the answer linking to which part of the context they're generated from? And I've found a great solution: a great technique called Layer-wise Relevance Propagation (LRP), showcased in a paper at ICML `24 by Reduan Achtibat et al allows, allows to precisely score how important each input token was in generating your output! They've made it into a library called LXT. πŸ“Š For each generated output token, LXT gives you attribution scores for each input token. βš™οΈ So I've worked a bit more on aggregating these scores into meaningful spans between successive input and output tokens, and I finally obtained my desired result: RAG with source highlighting! Try the demo here πŸ‘‰ https://huggingface.co/spaces/m-ric/rag_highlights Caveats: - It slows down generation (for now quite a lot, could hopefully be reduced a lot) - For now it supports only specific models: Llama models and Mixtral If there's enough interest in this solution, I can improve it further and spin it off into a specific library for RAG! πŸš€
View all activity

Organizations

Xdotnet's activity

Reacted to m-ric's post with πŸš€ about 2 months ago
view post
Post
1276
𝗔𝗱𝗱 π˜€π—Όπ˜‚π—Ώπ—°π—² π—΅π—Άπ—΄π—΅π—Ήπ—Άπ—΄π—΅π˜π—Άπ—»π—΄ π˜π—Ό π˜†π—Όπ˜‚π—Ώ π—₯π—”π—š π˜€π˜†π˜€π˜π—²π—Ί! πŸ“„πŸ’‘

RAG systems are supposed to make your LLM's answer more trustworthy, by inserting in the prompt some supporting documents from a knowledge base : we say that we're "adding some context".

πŸ‘Ž But if you don't know which part of the answer has been generated based on which input tokens, it's hard to tell wether it was effectively grounded in the context knowledge or not!

πŸ€” I've been working on the question: is it possible to add notes to the answer linking to which part of the context they're generated from?

And I've found a great solution: a great technique called Layer-wise Relevance Propagation (LRP), showcased in a paper at ICML `24 by Reduan Achtibat et al allows, allows to precisely score how important each input token was in generating your output! They've made it into a library called LXT.

πŸ“Š For each generated output token, LXT gives you attribution scores for each input token.

βš™οΈ So I've worked a bit more on aggregating these scores into meaningful spans between successive input and output tokens, and I finally obtained my desired result: RAG with source highlighting!

Try the demo here πŸ‘‰ m-ric/rag_highlights

Caveats:
- It slows down generation (for now quite a lot, could hopefully be reduced a lot)
- For now it supports only specific models: Llama models and Mixtral

If there's enough interest in this solution, I can improve it further and spin it off into a specific library for RAG! πŸš€
Reacted to davidberenstein1957's post with πŸ€— about 2 months ago
Reacted to victor's post with πŸ”₯ about 2 months ago
view post
Post
5450
πŸ™‹ Calling all Hugging Face users! We want to hear from YOU!

What feature or improvement would make the biggest impact on Hugging Face?

Whether it's the Hub, better documentation, new integrations, or something completely different – we're all ears!

Your feedback shapes the future of Hugging Face. Drop your ideas in the comments below! πŸ‘‡
Β·
Reacted to Jaward's post with πŸ‘ 4 months ago
view post
Post
1718
Super Exciting New Paper By MetaπŸ€–πŸ§ πŸš€

Discrete Flow Matching:
Introduces a new framework/algorithm for generating text/code without having to predict auto-regressively or one β€œword” at a time as traditional GPT models do. It generates all parts of the text/code at once.

The algorithm does this by slowly transforming random noise (source) into meaningful text (data). It learns how to transform samples along a path created between source and target using a "probability velocity" that describes how probabilities change over time. During generation, DFM starts with a random sample and iteratively updates it using this learned velocity, gradually transforming it into a sample from the target distribution. This allows for non-autoregressive generation.

They were able to scale models of up to 1.7B parameters achieving impressive scores on HumanEval and MBPP for coding, significantly closing the gap between autoregressive models and discrete flow models.

Though in its infancy, it sure does hold a promising future as leading research scientists argue non-autoregressive methods yield better reasoning.
Reacted to Symbol-LLM's post with πŸš€ 4 months ago
view post
Post
2116
πŸ”₯Thrilled to release our 8B version of Symbol-LLM-Instruct !

It follows the two-stage training strategy proposed in the original paper and is continually optimized on LLaMA3-Chat-8B model.

Symbol-LLM was accepted by ACL'24 main conference ! See you in Thailand !

Paper link: https://arxiv.org/abs/2311.09278
Paper Title: Symbol-LLM: Towards Foundational Symbol-centric Interface For Large Language Models
  • 1 reply
Β·
Reacted to TuringsSolutions's post with πŸ‘ 4 months ago
view post
Post
1379
SNN Image Diffusion V2

Billionaires have been made for less than this. This is only one of the things it can it do. It can do API calls, function calls, optimize poker and blackjack odds, anything that is an optimization problem. It costs fractions of a penny and requires fractions of the compute of an LLM model. It can even communicate two ways with an LLM model.
Β·
updated a model about 1 year ago
New activity in BRICS/README about 1 year ago

BRICS

#1 opened about 1 year ago by Xdotnet
liked a Space about 1 year ago
updated a Space about 1 year ago