Overview

Osmosis-MCP-4B is based on the Qwen3-4B model, fine-tuned with reinforcement learning to excel at multi step MCP-style tool usage.

We trained Osmosis-MCP-4B using a custom curriculum of multi-turn, tool-reliant prompts that mimic real-world use cases — for example:

"Given the weather in San Francisco, what are the top hiking locations?"

In addition, we provide a list of deterministic MCP like functions and mock server side behavior for the model to call and use.

This requires the model to reason through multiple tool invocations (e.g., weather → location ranker), and choose tools over intuition when applicable.


Training Approach

Our training pipeline leverages:

  • Dr. GRPO for stable and sample-efficient reinforcement learning.
  • Synthetic multi-step MCP interactions with strong tool chaining behavior, generated using our internal data engine.
  • SGLang + VeRL for efficient multi-turn rollout environments, built on top of Qwen3-4B for its function-calling capabilities.

Through this training methodology, we observed a notable behavioral shift: the model prefers invoking tools when appropriate, instead of relying solely on pre-trained intuition — a key milestone for MCP-native agents.


Why This Matters

MCP is fast becoming the open standard for tool-augmented AI agents. However:

  • Most top-performing models (e.g., Claude 3.7 Sonnet, Gemini 2.5 Pro) are closed.
  • Tool sprawl across clients and servers creates complexity.
  • Open models often lack the training to effectively use tools at all.

Osmosis-MCP-4B addresses all three — it’s small, powerful, and practical.


Downloads last month
156
Safetensors
Model size
4.41B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for osmosis-ai/osmosis-mcp-4b

Quantizations
2 models