GX Kok

gxkok
ยท

AI & ML interests

None yet

Recent Activity

reacted to singhsidhukuldeep's post with ๐Ÿง  15 days ago
O1 Embedder: Transforming Retrieval Models with Reasoning Capabilities Researchers from University of Science and Technology of China and Beijing Academy of Artificial Intelligence have developed a novel retrieval model that mimics the slow-thinking capabilities of reasoning-focused LLMs like OpenAI's O1 and DeepSeek's R1. Unlike traditional embedding models that directly match queries with documents, O1 Embedder first generates thoughtful reflections about the query before performing retrieval. This two-step process significantly improves performance on complex retrieval tasks, especially those requiring intensive reasoning or zero-shot generalization to new domains. The technical implementation is fascinating: - The model integrates two essential functions: Thinking and Embedding - It uses an "Exploration-Refinement" data synthesis workflow where initial thoughts are generated by an LLM and refined by a retrieval committee - A multi-task training method fine-tunes a pre-trained LLM to generate retrieval thoughts via behavior cloning while simultaneously learning embedding capabilities through contrastive learning - Memory-efficient joint training enables both tasks to share encoding results, dramatically increasing batch size The results are impressive - O1 Embedder outperforms existing methods across 12 datasets in both in-domain and out-of-domain scenarios. For example, it achieves a 3.9% improvement on Natural Questions and a 3.0% boost on HotPotQA compared to models without thinking capabilities. This approach represents a significant paradigm shift in retrieval technology, bridging the gap between traditional dense retrieval and the reasoning capabilities of large language models. What do you think about this approach? Could "thinking before retrieval" transform how we build search systems?
updated a model 18 days ago
123rf-data/flan-t5-xxl-fp16
published a model 18 days ago
123rf-data/flan-t5-xxl-fp16
View all activity

Organizations

123rf Data & AI's profile picture

gxkok's activity

reacted to singhsidhukuldeep's post with ๐Ÿง  15 days ago
view post
Post
3461
O1 Embedder: Transforming Retrieval Models with Reasoning Capabilities

Researchers from University of Science and Technology of China and Beijing Academy of Artificial Intelligence have developed a novel retrieval model that mimics the slow-thinking capabilities of reasoning-focused LLMs like OpenAI's O1 and DeepSeek's R1.

Unlike traditional embedding models that directly match queries with documents, O1 Embedder first generates thoughtful reflections about the query before performing retrieval. This two-step process significantly improves performance on complex retrieval tasks, especially those requiring intensive reasoning or zero-shot generalization to new domains.

The technical implementation is fascinating:

- The model integrates two essential functions: Thinking and Embedding
- It uses an "Exploration-Refinement" data synthesis workflow where initial thoughts are generated by an LLM and refined by a retrieval committee
- A multi-task training method fine-tunes a pre-trained LLM to generate retrieval thoughts via behavior cloning while simultaneously learning embedding capabilities through contrastive learning
- Memory-efficient joint training enables both tasks to share encoding results, dramatically increasing batch size

The results are impressive - O1 Embedder outperforms existing methods across 12 datasets in both in-domain and out-of-domain scenarios. For example, it achieves a 3.9% improvement on Natural Questions and a 3.0% boost on HotPotQA compared to models without thinking capabilities.

This approach represents a significant paradigm shift in retrieval technology, bridging the gap between traditional dense retrieval and the reasoning capabilities of large language models.

What do you think about this approach? Could "thinking before retrieval" transform how we build search systems?
reacted to MonsterMMORPG's post with ๐Ÿ”ฅ about 1 month ago
view post
Post
2791
MSI RTX 5090 TRIO FurMark Benchmarking + Overclocking + Noise Testing and Comparing with RTX 3090 TI

Video : https://youtu.be/uV3oqdILOmA

I have early purchased MSI RTX 5090 32G GAMING TRIO OC GPU to bring tests, benchmarks and experiments as I have promised. This is my introduction video to the GPU and first review. Hopefully many more amazing ones are coming very soon especially for Generative AI APPs and models. Moreover I am keeping my Gainward RTX 3090 Ti Phantom on the machine so that I can test and compare both GPUs.

In this video I will be FurMark testing both of the GPUs. Then I will overclock RTX 5090 and re-test. Then I will compare all test results. Furthermore I will do stress test of the both of the GPUs, checkout the temperatures and fan levels and I will compare their noise level.

๐Ÿ”— RTX 5090 Tests, Benchmarks, Reviews, Experiments Video Series Playlist โคต๏ธ
โ–ถ๏ธ https://www.youtube.com/playlist?list=PL_pbwdIyffslNd9aLizjQtHVHAMA6tpfT

0:00 Introduction to the RTX 5090 review
1:18 System specs are shown fully
1:33 Showing how FurMark runs at RTX 5090
2:40 MSI RTX 5090 32G GAMING TRIO OC FurMark benchmarking results
3:09 How to overclock RTX 5090 and FurMark benchmark scores with overclocked RTX 5090
3:35 Overclocked FurMark benchmark results of RTX 5090
4:00 How to run FurMark at your secondary GPU
4:29 How FurMark runs on RTX 3090 Ti
5:10 Comparison of all benchmarks as a graph
6:54 Noise testing of MSI RTX 5090 TRIO OC GPU when under 100% load via FurMark
7:51 Noise testing of Gainward RTX 3090 Ti Phantom GPU when under 100% load via FurMark

Check video description for more information : https://youtu.be/uV3oqdILOmA

Hopefully many more AI based benchmarks and tests coming so stay subscribed.
reacted to rwightman's post with โค๏ธ 6 months ago
view post
Post
2555
A 'small' MobileNet-V4 update, I just pushed weights for the smallest model I've trained in the series, a 0.5 width multiplier version of the MobileNet-V4 Conv Small.

Now you may look at this and say hey, why is this impressive? 64.8% top-1 and 2.2M params? MobileNetV3-Small 0.75, and MobileNet-V2 0.5 are both fewer params (at ~2M) and over 65% top-1, what gives? Well this is where MobileNet-V4 differs from the previous versions of the model family, it trades off (gives up) a little parameter efficiency for some computational efficiency.

So, let's look at the speed. On a 4090 w/ torchcompile
* 98K img/sec - timm/mobilenetv4_conv_small_050.e3000_r224_in1k
* 58K img/sec - timm/mobilenetv3_small_075.lamb_in1k
* 37K img/sec - timm/mobilenetv2_050.lamb_in1k

And there you go, if you have a need for speed, MNV4 is the better option.
reacted to merve's post with ๐Ÿค— 7 months ago
view post
Post
2277
amazing leaderboard by @rwightman , compare all the image backbones on various metrics against model performance

below is an example for top-k against inferred samples per second
timm/leaderboard
replied to m-ric's post 8 months ago
view reply

Perhaps using GPT-4o for evaluation is not the best way to do it?