AI Disruption

AI Disruption

Share this post

AI Disruption
AI Disruption
Baidu Open-Sources ERNIE 4.5 Series, Outperforms DeepSeek-V3

Baidu Open-Sources ERNIE 4.5 Series, Outperforms DeepSeek-V3

Baidu open-sources 10 ERNIE 4.5 models with MoE architecture. Features 424B parameters, multimodal capabilities, and outperforms competitors on benchmarks.

Meng Li's avatar
Meng Li
Jun 30, 2025
∙ Paid
8

Share this post

AI Disruption
AI Disruption
Baidu Open-Sources ERNIE 4.5 Series, Outperforms DeepSeek-V3
2
Share

"AI Disruption" Publication 7000 Subscriptions 20% Discount Offer Link.


Ernie 4.5 und X1: KI des chinesischen Baidu-Konzerns soll neue Maßstäbe  setzen - PC-WELT

Today, Baidu officially open-sourced the ERNIE 4.5 series models.

The ERNIE 4.5 series open-source models include 10 models in total, covering Mixture of Experts (MoE) models with activated parameter scales of 47B and 3B, respectively (with the largest model having a total parameter count of 424B), as well as a 0.3B dense parameter model. The pre-training weights and inference code are completely open-sourced.

Currently, the ERNIE 4.5 open-source series can be downloaded and deployed on platforms such as PaddlePaddle Galaxy Community and Hugging Face, while the open-source model API service is also available on Baidu Intelligent Cloud's Qianfan Large Model Platform.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Meng Li
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share