Sam Witteveen

samwit

AI & ML interests

conversational AI, NLU

Recent Activity

upvoted an article 4 months ago
Welcome, Gradio 5
liked a model 4 months ago
Mozilla/Llama-3.2-3B-Instruct-llamafile
updated a model 5 months ago
samwit/cot_gf_plat_2k_phi35_full
View all activity

Organizations

Red Dragon AI's profile picture Spaces-explorers's profile picture gg-tt's profile picture

samwit's activity

upvoted an article 4 months ago
reacted to akhaliq's post with 👍 12 months ago
view post
Post
In Search of Needles in a 10M Haystack

Recurrent Memory Finds What LLMs Miss

paper addresses the challenge of processing long documents using generative transformer models. To evaluate different approaches, we introduce BABILong, a new benchmark designed to assess model capabilities in extracting and processing distributed facts within extensive texts. Our evaluation, which includes benchmarks for GPT-4 and RAG, reveals that common methods are effective only for sequences up to 10^4 elements. In contrast, fine-tuning GPT-2 with recurrent memory augmentations enables it to handle tasks involving up to 10^7 elements. This achievement marks a substantial leap, as it is by far the longest input processed by any open neural network model to date, demonstrating a significant improvement in the processing capabilities for long sequences