Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

Kseniase 
posted an update 3 days ago
view post
Post
5730
10 awesome advanced LoRA approaches

Low-Rank Adaptation (LoRA) is the go-to method for efficient model fine-tuning that adds small low-rank matrices instead of retraining full models. The field isn’t standing still – new LoRA variants push the limits of efficiency, generalization, and personalization. So we’re sharing 10 of the latest LoRA approaches you should know about:

1. Mixture-of-LoRA-experts → Mixture of Low-Rank Adapter Experts in Generalizable Audio Deepfake Detection (2509.13878)
Adds multiple low-rank adapters (LoRA) into a model’s layers, and a routing mechanism activates the most suitable ones for each input. This lets the model adapt better to new unseen conditions

2. Amortized Bayesian Meta-Learning for LoRA (ABMLL) → Amortized Bayesian Meta-Learning for Low-Rank Adaptation of Large Language Models (2508.14285)
Balances global and task-specific parameters within a Bayesian framework to improve uncertainty calibration and generalization to new tasks without high memory or compute costs

3. AutoLoRA → AutoLoRA: Automatic LoRA Retrieval and Fine-Grained Gated Fusion for Text-to-Image Generation (2508.02107)
Automatically retrieves and dynamically aggregates public LoRAs for stronger T2I generation

4. aLoRA (Activated LoRA) → Activated LoRA: Fine-tuned LLMs for Intrinsics (2504.12397)
Only applies LoRA after invocation, letting the model reuse the base model’s KV cache instead of recomputing the full turn’s KV cache. Efficient in multi-turn conversations

5. LiLoRA (LoRA in LoRA) → LoRA in LoRA: Towards Parameter-Efficient Architecture Expansion for Continual Visual Instruction Tuning (2508.06202)
Shares the LoRA matrix A across tasks and additionally low-rank-decomposes matrix B to cut parameters in continual vision-text MLLMs

6. Sensitivity-LoRA → Sensitivity-LoRA: Low-Load Sensitivity-Based Fine-Tuning for Large Language Models (2509.09119)
Dynamically assigns ranks to weight matrices based on their sensitivity, measured using second-order derivatives

Read further below ↓
Also, subscribe to the Turing Post: https://www.turingpost.com/subscribe
  • 2 replies
·
prithivMLmods 
posted an update 2 days ago
view post
Post
4872
Dropping some experimental adapters for FLUX.1-Kontext-dev, including Photo-Restore-i2i, PhotoCleanser-i2i, Polaroid-Warm-i2i, Yarn-Photo-i2i, and Monochrome-Pencil. These were trained under various settings with minimal image pairs to achieve optimal results. The dataset result sets end pairs were synthesized using Gemini-2.5-Flash-Image-Preview and others.🤗✨

prithivMLmods/PhotoCleanser-i2i: Remove objects while preserving the rest of the image.
prithivMLmods/Photo-Restore-i2i: Restore old photos into moderately colorized, detailed images.
prithivMLmods/Polaroid-Warm-i2i: Seamless vintage Polaroid-style images with warm, faded tones.
prithivMLmods/Yarn-Photo-i2i: Convert images into yarn-stitched artwork while retaining key details.
prithivMLmods/Monochrome-Pencil: Turn images into monochrome pencil sketches while keeping original features.

✨Note: All the above models share the same auto-labeling multimodal VLM captioning model, prithivMLmods/DeepCaption-VLA-7B, which is used for refining edit instructions and accurately understanding attributions for the generations.

✨Collection: prithivMLmods/i2i-kontext-exp-68ce573b5c0623476b636ec7

.
.
.
To know more about it, visit the app page or the respective model page!!
merve 
posted an update 2 days ago
view post
Post
4725
large AI labs open-sourced a ton of models last week 🔥
here's few picks, find even more here merve/sep-16-releases-68d13ea4c547f02f95842f05 🤝
> IBM released a new Docling model with 258M params based on Granite (A2.0) 📝 ibm-granite/granite-docling-258M
> Xiaomi released 7B audio LM with base and instruct variants (MIT) XiaomiMiMo/mimo-audio-68cc7202692c27dae881cce0
> DecartAI released Lucy Edit, open Nano Banana 🍌 (NC) decart-ai/Lucy-Edit-Dev
> OpenGVLab released a family of agentic computer use models (3B/7B/32B) with the dataset 💻 OpenGVLab/scalecua-68c912cf56f7ff4c8e034003
> Meituan Longcat released thinking version of LongCat-Flash 💭 meituan-longcat/LongCat-Flash-Thinking
  • 1 reply
·
prithivMLmods 
posted an update 1 day ago
view post
Post
2217
Photo-Mate-i2i – a space for experimenting with adapters for image manipulation using Kontext adapters, including Photo-Restore-i2i, PhotoCleanser-i2i, Polaroid-Warm-i2i, Yarn-Photo-i2i, Monochrome-Pencil, and more. Try out the demo, and to learn more, visit the app page or the respective model pages!

⚡Demo: prithivMLmods/Photo-Mate-i2i
⚙️How to Use: prithivMLmods/Photo-Mate-i2i#2
👨‍🔧i2i-Kontext(Experimental LoRAs): prithivMLmods/i2i-kontext-exp-68ce573b5c0623476b636ec7

yeonseok-zeticai 
posted an update 1 day ago
view post
Post
2216
YOLOv11 Complete On-device Study
- {NPU vs GPU vs CPU} Across All Model Variants

We've just completed comprehensive benchmarking of the entire YOLOv11 family on ZETIC.MLange.
Here's what every ML engineer needs to know.

📊 Key Findings Across 5 Model Variants (XL to Nano):

1. NPU Dominance in Efficiency:
- YOLOv11n: 1.72ms on NPU vs 53.60ms on CPU (31x faster)
- Memory footprint: 0-65MB across all variants
- Consistent sub-10ms inference even on XL models

2. The Sweet Spot - YOLOv11s:
- NPU: 3.23ms @ 95.57% mAP
- Perfect balance: 36MB model, production-ready speed
- 10x faster than GPU, 30x faster than CPU

3. Surprising Discovery:
Medium models (YOLOv11m) show unusual GPU performance patterns - NPU outperforms GPU by 4x (9.55ms vs 35.82ms), suggesting current GPU kernels aren't optimized for mid-size architectures.

4. Production Insights:
- XL/Large: GPU still competitive for batch processing
- Small/Nano: NPU absolutely crushes everything else
- Memory scaling: Linear from 10MB (Nano) to 217MB (XL)
- Accuracy plateau: 95.5-95.7% mAP across S/M/L variants

Real-world Impact:
For edge deployment, YOLOv11s on NPU delivers server-level accuracy at embedded speeds. This changes everything for real-time applications.

🔗 Test these benchmarks yourself: https://mlange.zetic.ai/p/Steve/YOLOv11_comparison?tab=versions&version=5

📈 Full benchmark suite available now

The data speaks for itself.
NPUs aren't the future - they're the present for efficient inference.
Which variant fits your use case? Let's discuss in the comments.
Monica997 
posted an update 1 day ago
view post
Post
890
AI Just Made My Cat the King of Emojis 👑🐱😂

Never thought I’d see this — but with iMini’s nano banana model, my cat is now a full emoji + sticker pack 🎨✨

Used the 9-grid meme template + cartoon sticker generator, and in just ONE click 👉 my ordinary cat photo turned into a hilarious, cute, and super shareable set of stickers 💬🔥

No need to master complicated nano banana prompts — iMini handles everything.
Perfect for chats, socials, or just showing off your pet’s new “digital identity.”

👉 Try it here: https://imini.com/nano-banana

Who else wants their pet to be the next emoji star? 🌟
AdinaY 
posted an update about 23 hours ago
sergiopaniego 
posted an update 2 days ago
view post
Post
1058
This summer TRL leveled up for multimodal alignment 🌞

✅ New VLM alignment methods (MPO, GRPO, GSPO)
✅ Extended RLOO & Online DPO for VLMs
✅ Native SFT support
✅ Ready-to-use training scripts

🔗 https://huggingface.co/blog/trl-vlm-alignment
yeonseok-zeticai 
posted an update about 5 hours ago
view post
Post
215
🎯 RetinaFace On-Device Deployment Study: NPU Acceleration Breakthrough!
(Check details at :https://mlange.zetic.ai/p/Steve/RetinaFace)

TL;DR: Successfully deployed RetinaFace with ZETIC.MLange achieving 1.43ms inference on mobile NPU!

🔍 Complete Performance Analysis:
Latency Comparison:
- NPU: 1.43ms (Winner! 🏆)
- GPU: 3.75ms
- CPU: 21.42ms

Accuracy Metrics - SNR:
- FP16: 56.98 dB
- Integer Quantized: 48.03 dB
(Precision-Performance: Excellent trade-off maintained)

Memory Footprint:
- Model Size: 2.00 MB (highly compressed)
- Runtime Memory: 14.58 MB peak
- Deployment Ready: ✅ Production optimized

🛠 Technical Implementation:
(Runnable with Copy & Paste at the MLange link!)

📊 Device Compatibility Matrix:
Tested on 50+ devices including Samsung Galaxy series, Google Pixel lineup, and Xiaomi devices, iPhones and iPads.
Consistent sub-5ms performance across the board!

🚀 Applications Unlocked:
- Real-time AR/VR face tracking
- Privacy-preserving edge authentication
- Live video processing pipelines
- Mobile security applications
- Interactive camera filters

The democratization of high-performance computer vision on mobile devices is happening NOW! This study proves that complex CV models can run efficiently on consumer hardware without compromising accuracy.
Want to reproduce these results? Check out the benchmark methodology and implementation guide!
kanaria007 
posted an update 1 day ago
view post
Post
938
✅ New Article: *Epilogue — The Structured Intelligence Computer*

Title:
🖥️ SIC-Epilogue: The Structured Intelligence Computer
🔗 https://huggingface.co/blog/kanaria007/sic-epilogue

---

Summary:
With *SPU* (processing), *GSPU* (simulation), and *SIM/SIS* (memory & storage),
the architecture of the *Structured Intelligence Computer (SIC)* is now complete.

This epilogue unifies the pieces into one design,
and shows why even peripherals like an *AmuSphere-class interface* become inevitable extensions once SIC exists.

> The point is not BIC vs AmuSphere.
> *It is structured, safe, explainable immersion under one law of intelligence.*

---

Why It Matters:
• Marks the closure of the hardware PoC arc
• Shows the computer as not only faster, but *resilient, ethical, and auditable*
• Positions SIC as the substrate for AGI, education, and exploration

---

What’s Inside:
• Integration of SPU, GSPU, SIM/SIS into a single SIC stack
• Peripheral emergence (AmuSphere-class) as natural consequence
• Reflection on why openness and safety matter at hardware level
• How this closes one arc but sets up the next

---

📖 PoC Series — Article 7

This article concludes the *hardware PoC arc*.
With the foundation in place, the series now shifts focus:
from hardware prototypes to *multi-domain combinations*,
demonstrating how structural protocols bridge fields once thought unrelated.

---

Next: *Acting & Self-Reference — A Unified Structural Failure*
The following article begins the *multi-domain combination series*,
showing how an actor lost in a role and an AI lost in recursion mirror the same structural problem.

> From machines to minds,
> *structure reveals its unity across domains.*