Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
1
1
2
Shetu Mohanto
shetumohanto
Follow
0 followers
ยท
28 following
ShetuMohanto
shetumohanto
shetumohanto
AI & ML interests
GenAI | MLOps | AI agent | Computer Vision
Recent Activity
reacted
to
clem
's
post
with ๐
9 days ago
What are the best organizations to follow on @huggingface? On top of my head: - Deepseek (35,000 followers): https://huggingface.co/deepseek-ai - Meta Llama (27,000 followers): https://huggingface.co/meta-llama - Black Forrest Labs (11,000 followers): https://huggingface.co/black-forest-labs - OpenAI (5,000 followers): https://huggingface.co/openai - Nvidia (16,000 followers): https://huggingface.co/nvidia - MIcrosoft (9,000 followers): https://huggingface.co/microsoft - AllenAI (2,000 followers): https://huggingface.co/allenai - Mistral (5,000 followers): https://huggingface.co/mistralai - XAI (600 followers): https://huggingface.co/xai-org - Stability AI (16,000 followers): https://huggingface.co/stabilityai - Qwen (16,000 followers): https://huggingface.co/Qwen - GoogleAI (8,000 followers): https://huggingface.co/google - Unsloth (3,000 followers): https://huggingface.co/unsloth - Bria AI (4,000 followers): https://huggingface.co/briaai - NousResearch (1,300 followers): https://huggingface.co/NousResearch Bonus, the agent course org with 17,000 followers: https://huggingface.co/agents-course
reacted
to
clem
's
post
with ๐ฅ
9 days ago
What are the best organizations to follow on @huggingface? On top of my head: - Deepseek (35,000 followers): https://huggingface.co/deepseek-ai - Meta Llama (27,000 followers): https://huggingface.co/meta-llama - Black Forrest Labs (11,000 followers): https://huggingface.co/black-forest-labs - OpenAI (5,000 followers): https://huggingface.co/openai - Nvidia (16,000 followers): https://huggingface.co/nvidia - MIcrosoft (9,000 followers): https://huggingface.co/microsoft - AllenAI (2,000 followers): https://huggingface.co/allenai - Mistral (5,000 followers): https://huggingface.co/mistralai - XAI (600 followers): https://huggingface.co/xai-org - Stability AI (16,000 followers): https://huggingface.co/stabilityai - Qwen (16,000 followers): https://huggingface.co/Qwen - GoogleAI (8,000 followers): https://huggingface.co/google - Unsloth (3,000 followers): https://huggingface.co/unsloth - Bria AI (4,000 followers): https://huggingface.co/briaai - NousResearch (1,300 followers): https://huggingface.co/NousResearch Bonus, the agent course org with 17,000 followers: https://huggingface.co/agents-course
reacted
to
mmhamdy
's
post
with โค๏ธ
9 days ago
๐ We're excited to introduce MemoryCode, a novel synthetic dataset designed to rigorously evaluate LLMs' ability to track and execute coding instructions across multiple sessions. MemoryCode simulates realistic workplace scenarios where a mentee (the LLM) receives coding instructions from a mentor amidst a stream of both relevant and irrelevant information. ๐ก But what makes MemoryCode unique?! The combination of the following: โ Multi-Session Dialogue Histories: MemoryCode consists of chronological sequences of dialogues between a mentor and a mentee, mirroring real-world interactions between coworkers. โ Interspersed Irrelevant Information: Critical instructions are deliberately interspersed with unrelated content, replicating the information overload common in office environments. โ Instruction Updates: Coding rules and conventions can be updated multiple times throughout the dialogue history, requiring LLMs to track and apply the most recent information. โ Prospective Memory: Unlike previous datasets that cue information retrieval, MemoryCode requires LLMs to spontaneously recall and apply relevant instructions without explicit prompts. โ Practical Task Execution: LLMs are evaluated on their ability to use the retrieved information to perform practical coding tasks, bridging the gap between information recall and real-world application. ๐ Our Findings 1๏ธโฃ While even small models can handle isolated coding instructions, the performance of top-tier models like GPT-4o dramatically deteriorates when instructions are spread across multiple sessions. 2๏ธโฃ This performance drop isn't simply due to the length of the context. Our analysis indicates that LLMs struggle to reason compositionally over sequences of instructions and updates. They have difficulty keeping track of which instructions are current and how to apply them. ๐ Paper: https://huggingface.co/papers/2502.13791 ๐ฆ Code: https://github.com/for-ai/MemoryCode
View all activity
Organizations
shetumohanto
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
6 months ago
mattshumer/Reflection-Llama-3.1-70B
Text Generation
โข
Updated
Sep 24, 2024
โข
608
โข
1.71k
liked
a model
10 months ago
mistralai/Mistral-7B-v0.1
Text Generation
โข
Updated
Jul 24, 2024
โข
372k
โข
โข
3.63k