AI Disruption

AI Disruption

Share this post

AI Disruption
AI Disruption
200B Model Beats DeepSeek-R1 – ByteDance Launches Seed-Thinking-v1.5
Copy link
Facebook
Email
Notes
More

200B Model Beats DeepSeek-R1 – ByteDance Launches Seed-Thinking-v1.5

ByteDance's Seed-Thinking-v1.5: 200B MoE model beats DeepSeek-R1 in reasoning & STEM tasks. Outperforms in benchmarks like AIME & Codeforces.

Meng Li's avatar
Meng Li
Apr 11, 2025
∙ Paid
2

Share this post

AI Disruption
AI Disruption
200B Model Beats DeepSeek-R1 – ByteDance Launches Seed-Thinking-v1.5
Copy link
Facebook
Email
Notes
More
3
Share

"AI Disruption" Publication 5800 Subscriptions 20% Discount Offer Link.


The ByteDance Doubao team today released a technical report for their new inference model Seed-Thinking-v1.5.

According to the report, this is a Mixture-of-Experts (MoE) model with a total of 200 billion parameters, activating 20 billion parameters during operation.

Its performance is remarkably impressive, surpassing DeepSeek-R1, which has 671 billion total parameters across various benchmarks.

Some speculate that this is the deep reasoning model currently in use by ByteDance’s Doubao.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Meng Li
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More