AI Disruption

AI Disruption

Share this post

AI Disruption
AI Disruption
DeepSeek/o3 Weakness: Why the Shortest Answer is Often Correct
Copy link
Facebook
Email
Notes
More

DeepSeek/o3 Weakness: Why the Shortest Answer is Often Correct

Recent research highlights the "Underthinking" issue in AI models like DeepSeek, showing how ineffective switching between approaches can waste resources and reduce accuracy.

Meng Li's avatar
Meng Li
Feb 04, 2025
∙ Paid
1

Share this post

AI Disruption
AI Disruption
DeepSeek/o3 Weakness: Why the Shortest Answer is Often Correct
Copy link
Facebook
Email
Notes
More
1
Share

"AI Disruption" publication New Year 30% discount link.


As DeepSeek and other inference models like o1/o3 continue to make a huge impact, some have started to study their weaknesses.

Recent research reveals:

When faced with difficult problems, inference models may frequently switch between different approaches like a “fickle student,” but fail due to lack of in-depth exploration. This phenomenon is referred to by researchers as Underthinking.

The research team, from Tencent AI Lab, Soochow University, and Shanghai Jiao Tong University, focused on the open-source DeepSeek-R1 and Qwen QwQ series models.

By analyzing AI's incorrect answers, they discovered that these models often start off on the right track early in their reasoning but tend to “scratch the surface” before quickly exploring other approaches. This leads to thousands of generated tokens that do not contribute to solving the problem.

This “ineffective effort” not only wastes computational resources but significantly reduces the accuracy of the answers.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Meng Li
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More