Rutam Risaldar

thugCodeNinja
·

AI & ML interests

Machine Learning, Deep Learning, NLP, LLM, Explainable AI

Recent Activity

liked a model 16 days ago
thugCodeNinja/robertafinetune
reacted to singhsidhukuldeep's post with 🚀 2 months ago
Groundbreaking Research Alert: Rethinking RAG with Cache-Augmented Generation (CAG) Researchers from National Chengchi University and Academia Sinica have introduced a paradigm-shifting approach that challenges the conventional wisdom of Retrieval-Augmented Generation (RAG). Instead of the traditional retrieve-then-generate pipeline, their innovative Cache-Augmented Generation (CAG) framework preloads documents and precomputes key-value caches, eliminating the need for real-time retrieval during inference. Technical Deep Dive: - CAG preloads external knowledge and precomputes KV caches, storing them for future use - The system processes documents only once, regardless of subsequent query volume - During inference, it loads the precomputed cache alongside user queries, enabling rapid response generation - The cache reset mechanism allows efficient handling of multiple inference sessions through strategic token truncation Performance Highlights: - Achieved superior BERTScore metrics compared to both sparse and dense retrieval RAG systems - Demonstrated up to 40x faster generation times compared to traditional approaches - Particularly effective with both SQuAD and HotPotQA datasets, showing robust performance across different knowledge tasks Why This Matters: The approach significantly reduces system complexity, eliminates retrieval latency, and mitigates common RAG pipeline errors. As LLMs continue evolving with expanded context windows, this methodology becomes increasingly relevant for knowledge-intensive applications.
View all activity

Organizations

None yet

thugCodeNinja's activity

New activity in thugCodeNinja/robertafinetune 16 days ago
reacted to singhsidhukuldeep's post with 🚀 2 months ago
view post
Post
3208
Groundbreaking Research Alert: Rethinking RAG with Cache-Augmented Generation (CAG)

Researchers from National Chengchi University and Academia Sinica have introduced a paradigm-shifting approach that challenges the conventional wisdom of Retrieval-Augmented Generation (RAG).

Instead of the traditional retrieve-then-generate pipeline, their innovative Cache-Augmented Generation (CAG) framework preloads documents and precomputes key-value caches, eliminating the need for real-time retrieval during inference.

Technical Deep Dive:
- CAG preloads external knowledge and precomputes KV caches, storing them for future use
- The system processes documents only once, regardless of subsequent query volume
- During inference, it loads the precomputed cache alongside user queries, enabling rapid response generation
- The cache reset mechanism allows efficient handling of multiple inference sessions through strategic token truncation

Performance Highlights:
- Achieved superior BERTScore metrics compared to both sparse and dense retrieval RAG systems
- Demonstrated up to 40x faster generation times compared to traditional approaches
- Particularly effective with both SQuAD and HotPotQA datasets, showing robust performance across different knowledge tasks

Why This Matters:
The approach significantly reduces system complexity, eliminates retrieval latency, and mitigates common RAG pipeline errors. As LLMs continue evolving with expanded context windows, this methodology becomes increasingly relevant for knowledge-intensive applications.
updated a Space 4 months ago
updated a Space 8 months ago
updated a Space 10 months ago
liked a Space 11 months ago
updated a Space 12 months ago
liked a Space 12 months ago
updated a Space 12 months ago
liked a Space 12 months ago
updated a Space about 1 year ago
updated a Space about 1 year ago