Baidu releases ERNIE-5.1-Preview, achieving top LMArena rankings with 1/3 parameter efficiency and advanced agent capabilities.

Baidu has officially unveiled ERNIE-5.1-Preview, a significant leap in large language model efficiency and performance released on May 9, 2026. This model addresses the industry's growing need for cost-effective inference without sacrificing intelligence. Unlike previous iterations that relied on brute-force scaling, ERNIE-5.1-Preview introduces a highly optimized architecture designed for enterprise-grade applications.
The release marks a strategic pivot for Baidu, positioning their model to compete directly with global leaders like DeepSeek and OpenAI's latest offerings. By focusing on parameter efficiency and specialized post-training, Baidu aims to democratize access to high-quality AI capabilities for developers and businesses operating within the Chinese market and globally.
The core innovation of ERNIE-5.1-Preview lies in its Mixture of Experts (MoE) architecture. Baidu has successfully compressed the total parameters to approximately one-third of the previous ERNIE-5.0 model, while active parameters are reduced to roughly half. This drastic reduction is achieved without compromising the model's ability to process complex tasks, demonstrating advanced sparse training techniques.
Technically, the model is powered by a decoupled fully-asynchronous reinforcement learning infrastructure built on Baidu's PaddlePaddle framework. It utilizes a scaled agentic post-training pipeline featuring Multi-Teacher On-Policy Distillation (MOPD). This four-stage pipeline includes Supervised Fine-Tuning (SFT), Domain Expert Model Training, On-Policy Distillation, and General Online RL, ensuring robust knowledge retention.
ERNIE-5.1-Preview has demonstrated exceptional performance across specialized leaderboards, topping the LMArena Search leaderboard as #4 globally and #1 among Chinese models with a score of 1,223. Its specialized capabilities are particularly noteworthy, securing the #1 global rank in Legal & Government and #4 in Business, Management & Financial Ops.
In terms of reasoning and coding, the model achieves a score of 99.6 on AIME26 with tool use, ranking second only to Gemini 3.1 Pro. It also surpasses DeepSeek-V4-Pro on critical agent evaluation tasks like tau-cubed-bench and SpreadsheetBench-Verified. This performance is achieved at a fraction of the pre-training cost, highlighting the efficiency of the underlying architecture.
Baidu has not yet published specific public token pricing for ERNIE-5.1-Preview on their standard API documentation page. Pricing is typically negotiated on an enterprise basis or varies by region. Developers should contact Baidu Cloud sales for current rate cards.
While specific input and output costs per million tokens are currently unavailable, the model's 6% pre-training cost efficiency suggests a competitive operational expenditure for inference. The model is currently rolling out on internal creative production agent platforms, indicating a focus on enterprise integration rather than immediate public API consumption.
ERNIE-5.1-Preview is best suited for high-stakes domains requiring precision and compliance. Its dominance in Legal & Government leaderboards makes it ideal for contract analysis, regulatory compliance checks, and government document processing. The model's ability to handle complex reasoning tasks positions it well for financial operations and risk management systems.
Beyond traditional text generation, the model is being deployed on creative production agent platforms including ISEKAI ZERO and Mulan AI. Its creative writing capabilities approach those of Gemini 3.1 Pro, making it suitable for content generation, game development, and marketing automation where high-quality narrative coherence is essential.
Access to ERNIE-5.1-Preview is currently limited to specific production platforms and enterprise partners. Developers interested in integration should monitor the Baidu ERNIE official blog for public API endpoints. The model is currently rolling out on 10+ creative production agent platforms.
For those looking to experiment with the architecture, Baidu provides documentation on their official blog regarding the release. Developers should prepare for integration by reviewing the four-stage post-training pipeline details to understand the model's knowledge boundaries and optimal prompting strategies for the MoE structure.