Michael Fried

MichaelFried

AI & ML interests

None yet

Recent Activity

liked a model 5 months ago
juliozhao/DocLayout-YOLO-DocStructBench
liked a model 5 months ago
PleIAs/OCRonos
liked a dataset 5 months ago
Berzerker/incidental_scene_ocr_dataset
View all activity

Organizations

None yet

MichaelFried's activity

reacted to singhsidhukuldeep's post with 👍 8 months ago
view post
Post
2347
Hello, HuggingFace🤗 community 🌟,

All the amazing people quantising LLMs to AWQ and GPTQ 🔧🤖

Can you please mention the perplexity you achieved 📉 OR any other metric to measure the quantisation qualitatively? 📊

The GGUF community follows this really well! 👍

And if it is not too much to ask, the script used for quantisation would be amazing! 📝

Thanks for the quants for the GPU poor! 💻
  • 1 reply
·
reacted to osanseviero's post with ❤️ 12 months ago
view post
Post
Diaries of Open Source. Part 4!

🌏Cohere and Cohere4AI release Command-R, a 35B model that is multilingual, RAG-optimized, and can manage tools!
Model: CohereForAI/c4ai-command-r-v01
Blog post: https://txt.cohere.com/command-r/

🧑‍🍳StarChat2: A powerful code model that is conversational
Try it out: HuggingFaceH4/starchat2-playground
Repos: HuggingFaceH4/starchat2-15b-65f068417b330fafad751fce
Training code: https://github.com/huggingface/alignment-handbook/tree/main/recipes/starchat2-15b

🐲Yi-9B: trained on 3 trillion tokens, this english-chinese LLM is quite good and with a very nice detailed report!
Model: 01-ai/Yi-9B
Paper: Yi: Open Foundation Models by 01.AI (2403.04652)

🐋DeepSeek-VL, 1.3B and 7B VLMs
Paper: DeepSeek-VL: Towards Real-World Vision-Language Understanding (2403.05525)
Large model: deepseek-ai/deepseek-vl-7b-chat

✍️Writer releases OmniACT: a dataset for multimodal agents for desktop and web.
Dataset: Writer/omniact
Paper: OmniACT: A Dataset and Benchmark for Enabling Multimodal Generalist Autonomous Agents for Desktop and Web (2402.17553)

🍎Apple releases MobileCLIP: fast image-text models! https://github.com/apple/ml-mobileclip

🦙💪LlamaGym - fine-tune LLM agents with RL in just a few lines of code! https://github.com/KhoomeiK/LlamaGym

🖼️New multimodal leaderboard ConTextual https://huggingface.co/blog/leaderboard-contextual

🎁 Design2Code: benchmark for multimodal LLMs for automating front-end development.
Dataset SALT-NLP/Design2Code
Paper Design2Code: How Far Are We From Automating Front-End Engineering? (2403.03163)
Project https://salt-nlp.github.io/Design2Code/

You can find the previous part at https://huggingface.co/posts/osanseviero/633758457910104