The bottom line: StepFun’s Step3-VL-10B shatters the bigger-is-better myth by outperforming proprietary models twenty times its size. This open-source powerhouse uses smart parallel reasoning to deliver elite multimodal capabilities without the massive hardware costs. With an impressive 94.43% score on AIME 2025, it proves efficient AI can still be state-of-the-art.
Think you need massive servers for top-tier AI? Think again. A new underdog just arrived, and it is punching way above its weight class. We dive into Step3 VL 10B, the open-source marvel outperforming giants ten times its size, to see how it achieves the impossible.
StepFun’s New Model: Small Size, Massive Performance

The AI race usually favors the gigantic, but a new contender just flipped the script on the “bigger is better” obsession.
What Exactly Is Step3-VL-10B?
StepFun dropped Step3-VL-10B, and it’s not just another model. It is a multimodal beast, processing text and images seamlessly. Best of all? They made this tech fully open-source, breaking down the usual walled gardens.
Here is the kicker: it runs on just 10 billion parameters. That’s incredibly light. StepFun deliberately engineered this to hit the sweet spot between raw efficiency and high-level intelligence.
You can grab it right now. It operates under the Apache 2.0 license, available freely on platforms like Hugging Face and ModelScope.
Punching Way Above Its Weight Class
The real story isn’t the size; it’s the output. This model doesn’t just compete; it rivals and sometimes beats systems that are 10 to 20 times larger.
We are talking about heavy hitters like GLM-4.6V and Qwen3-VL-Thinking. Seeing a compact model stand toe-to-toe with these giants is frankly startling for the industry.
Despite its compact size, Step3-VL-10B achieves state-of-the-art results, even outperforming leading proprietary models like Gemini 2.5 Pro and Seed-1.5-VL on several key benchmarks.
This isn’t luck. It stems from highly specific, intentional design choices in the architecture.
The Engineering Behind the Breakthrough
So, how does a team manage to create a model so small yet so powerful? The answer lies in their approach to training and a clever inference trick.
A Smarter, Unified Training Approach
Most labs build these things piece by piece, hoping they stick. StepFun took a different route. They trained the step3 vl 10b model in one go—fully unfrozen—on a massive dataset of 1.2 trillion tokens. It’s a bold move that pays off.
- Unified Pre-training: The visual and language parts (the PE-lang encoder and Qwen3-8B decoder) were trained simultaneously for better synergy.
- High-Quality Data Focus: The training data was curated to target complex perception (like OCR and GUI interaction) and general reasoning tasks.
- Advanced Fine-Tuning: The model underwent over 1,400 iterations of reinforcement learning (RLVR and RLHF) to sharpen its advanced abilities.
If you want to see the math behind the magic, check the official technical report. It details exactly why this unified method beats the standard fragmented approach.
PaCoRe: The Parallel Reasoning Trick That Changes Everything
Here is the real secret sauce. It’s called PaCoRe, or Parallel Coordinated Reasoning. This isn’t about how the model learns; it’s about how it “thinks” when you ask it a tough question. It fundamentally changes the output quality.
Instead of relying on a single chain of thought, PaCoRe launches 16 parallel explorations of an image. It gathers evidence from all angles before synthesizing the final answer.
| Benchmark | Standard Mode (SeRe) | Advanced Mode (PaCoRe) | Performance Gain |
|---|---|---|---|
| AIME 2025 | High | 94.43% | Significant Boost |
| MathVision | High | 75.95% | Significant Boost |
| Context Window | 64K tokens | 128K tokens | Doubled Capacity |
This method is why the model crushes complex reasoning tasks that usually stump significantly larger architectures.
What This Model Can Actually Do (and Where to Find It)
Specs on paper are fine, but let’s get real: what can this thing actually do for you? And more importantly, how do you get your hands on it?
From Complex Math to Reading Screens
Let’s skip the fluff and look at the raw capabilities. Here is exactly where the Step3-VL-10B flexes its muscles.
- STEM Reasoning: Achieves an impressive 94.43% on AIME 2025 and 75.95% on MathVision, showcasing elite math and science capabilities.
- Visual Perception: Scores 92.05% on MMBench (EN), proving its strong general visual understanding.
- GUI & OCR: Excels at reading interfaces and text in images, with a score of 86.75% on OCRBench.
- Coding Ability: Demonstrates solid coding skills with a 66.05% score on HumanEval-V.
These aren’t just vanity metrics. They translate to serious horsepower for tasks ranging from automated software debugging to complex educational tutoring systems that actually work.
Open Source for Everyone: How to Get Started
Here is the kicker: it’s fully open source under the permissive Apache 2.0 license. That means you can grab it, tweak it, and build whatever you want without corporate lawyers breathing down your neck.
StepFun’s decision to open-source the model and its weights on platforms like Hugging Face and ModelScope accelerates community-driven progress and makes powerful AI more accessible to all.
The dev scene is already buzzing about this release. You can pull the weights right now from the official GitHub repository. Since it plugs directly into the Transformers library, you don’t need a PhD to get it running locally.
And no, this has absolutely nothing to do with the “Step 3” medical exams—we’re strictly talking silicon intelligence here.
Step3-VL-10B proves that in the AI world, size isn’t everything. With its smart engineering and that clever PaCoRe trick, this open-source gem punches way above its weight class. It is definitely worth checking out for your next project. Who knew 10 billion parameters could be this smart?





