My (Chiffon) Nguyen's picture
7 9

My (Chiffon) Nguyen PRO

chiffonng
·

AI & ML interests

human-centric and data-efficient AI for knowledge acquisition

Recent Activity

Organizations

None yet

chiffonng's activity

upvoted an article 5 days ago
view article
Article

SmolLM - blazingly fast and remarkably powerful

336
upvoted an article 6 days ago
view article
Article

Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM

310
upvoted an article 9 days ago
view article
Article

Fixing Gradient Accumulation

51
upvoted an article 18 days ago
view article
Article

SmolVLM - small yet mighty Vision Language Model

222
reacted to singhsidhukuldeep's post with 🧠 19 days ago
view post
Post
3476
O1 Embedder: Transforming Retrieval Models with Reasoning Capabilities

Researchers from University of Science and Technology of China and Beijing Academy of Artificial Intelligence have developed a novel retrieval model that mimics the slow-thinking capabilities of reasoning-focused LLMs like OpenAI's O1 and DeepSeek's R1.

Unlike traditional embedding models that directly match queries with documents, O1 Embedder first generates thoughtful reflections about the query before performing retrieval. This two-step process significantly improves performance on complex retrieval tasks, especially those requiring intensive reasoning or zero-shot generalization to new domains.

The technical implementation is fascinating:

- The model integrates two essential functions: Thinking and Embedding
- It uses an "Exploration-Refinement" data synthesis workflow where initial thoughts are generated by an LLM and refined by a retrieval committee
- A multi-task training method fine-tunes a pre-trained LLM to generate retrieval thoughts via behavior cloning while simultaneously learning embedding capabilities through contrastive learning
- Memory-efficient joint training enables both tasks to share encoding results, dramatically increasing batch size

The results are impressive - O1 Embedder outperforms existing methods across 12 datasets in both in-domain and out-of-domain scenarios. For example, it achieves a 3.9% improvement on Natural Questions and a 3.0% boost on HotPotQA compared to models without thinking capabilities.

This approach represents a significant paradigm shift in retrieval technology, bridging the gap between traditional dense retrieval and the reasoning capabilities of large language models.

What do you think about this approach? Could "thinking before retrieval" transform how we build search systems?