{"id":224347,"date":"2026-01-06T23:33:09","date_gmt":"2026-01-06T23:33:09","guid":{"rendered":"https:\/\/www.newsbeep.com\/il\/224347\/"},"modified":"2026-01-06T23:33:09","modified_gmt":"2026-01-06T23:33:09","slug":"deepseek-v3-2-outperforms-gpt-5-on-reasoning-tasks","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/il\/224347\/","title":{"rendered":"DeepSeek-V3.2 Outperforms GPT-5 on Reasoning Tasks"},"content":{"rendered":"<p><a href=\"https:\/\/www.deepseek.com\/en\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek<\/a> released <a href=\"https:\/\/api-docs.deepseek.com\/news\/news251201\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek-V3.2<\/a>, a family of open-source reasoning and agentic AI models. The high compute version, DeepSeek-V3.2-Speciale, performs better than GPT-5 and comparably to Gemini-3.0-Pro on several reasoning benchmarks.<\/p>\n<p>DeepSeek applied three new techniques in the development of DeepSeek-V3.2. First, they used a more efficient attention mechanism called DeepSeek Sparse Attention (DSA) that reduces the computational complexity of the model. They also scaled the reinforcement learning phase, which consumed more compute budget than did pre-training. Finally, they developed an agentic task synthesis pipeline to improve the models&#8217; tool use. The result was a model that outperforms most other open models on a range of coding, reasoning, and agentic benchmarks, and performs as well as or better than closed frontier models such as GPT-5 and Gemini-3.0-Pro. However, the DeepSeek team pointed out:<\/p>\n<p>&#13;<\/p>\n<p>Despite these achievements, we acknowledge certain limitations when compared to frontier closed-source models&#8230;First, due to fewer total training FLOPs, the breadth of world knowledge in DeepSeek-V3.2 still lags behind that of leading proprietary models. We plan to address this knowledge gap in future iterations by scaling up the pre-training compute. Second, token efficiency remains a challenge&#8230;Future work will focus on optimizing the intelligence density of the model\u2019s reasoning chains to improve efficiency. Third, solving complex tasks is still inferior to frontier models, motivating us to further refine our foundation model and post-training recipe.<\/p>\n<p>&#13;<\/p>\n<p>InfoQ covered several of DeepSeek&#8217;s previous releases, including the initial <a href=\"https:\/\/www.infoq.com\/news\/2025\/01\/deepseek-v3-llm\/\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek-V3<\/a> launch and <a href=\"https:\/\/www.infoq.com\/news\/2025\/02\/deepseek-r1-release\/\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek-R1<\/a>, their first reasoning model; both were released in early 2025. Later in 2025, InfoQ covered <a href=\"https:\/\/www.infoq.com\/news\/2025\/09\/deepseek-v31-hybrid\/\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek-V3.1<\/a>, a hybrid reasoning model that combines thinking and non-thinking modes in a single system.<\/p>\n<p><img decoding=\"async\" alt=\"DeepSeek-V3.2 Benchmark Performance\" style=\"width: 1000px; height: 548px;\" src=\"https:\/\/www.infoq.com\/news\/2026\/01\/deepseek-v32\/news\/2026\/01\/deepseek-v32\/en\/resources\/1DeepSeek-V3.2-Benchmarks-1767532715071.png\" rel=\"share\"\/><\/p>\n<p>DeepSeek-V3.2 Benchmark Performance. Image Source: <a href=\"https:\/\/huggingface.co\/deepseek-ai\/DeepSeek-V3.2\/resolve\/main\/assets\/paper.pdf\" rel=\"nofollow noopener\" target=\"_blank\">DeepSeek Tech Report<\/a><\/p>\n<p>DeepSeek-V3.2 uses the same architecture as DeepSeek-V3.1, except using the new DSA attention mechanism. The team started with a checkpoint of DeepSeek-V3.1 and extended the context length to 128K before continuing pre-training to produce DeepSeek-V3.2. The new attention mechanism reduces the computational complexity from O(^2) to O(), where L is context length and k&lt;&#13;<br \/>\n&#13;<\/p>\n<p>For post-training, the team used specialist distillation. They trained a set of specialist models dedicated to a particular domain: coding, math, and several agent tasks. Then these specialist models produce synthetic training data that is used to fine-tune the main model.<\/p>\n<p>&#13;<br \/>\n&#13;<\/p>\n<p>In a Hacker News <a href=\"https:\/\/news.ycombinator.com\/item?id=46108780\" rel=\"nofollow noopener\" target=\"_blank\">discussion about DeepSeek-V3.2<\/a>, several users pointed out the advantages of a high-performing open model. One user wrote:<\/p>\n<p>&#13;<br \/>\n&#13;<br \/>\n&#13;<\/p>\n<p>If you&#8217;re trying to build AI based applications you can and should compare the costs between vendor based solutions and hosting open models with your own hardware&#8230;Then you compare that to the cost of something like GPT-5, which is a bit simpler because the cost per (million) token is something you can grab off of a website. You&#8217;d be surprised how much money running something like DeepSeek (or if you prefer a more established company, Qwen3) will save you over the cloud systems&#8230;DeepSeek and Qwen will function on cheap GPUs that other models will simply choke on.<\/p>\n<p>&#13;<br \/>\n&#13;<br \/>\n&#13;<\/p>\n<p>The DeepSeek-V3.2 model files are <a href=\"https:\/\/huggingface.co\/deepseek-ai\/DeepSeek-V3.2\" rel=\"nofollow noopener\" target=\"_blank\">available to download<\/a> from Huggingface. However, the high-compute DeepSeek-V3.2-Speciale variant is currently only available via DeepSeek&#8217;s API.<\/p>\n<p>&#13;<\/p>\n<p>&#13;<br \/>\n&#13;<br \/>\n&#13;<br \/>\n&#13;<br \/>\n&#13;<\/p>\n","protected":false},"excerpt":{"rendered":"DeepSeek released DeepSeek-V3.2, a family of open-source reasoning and agentic AI models. The high compute version, DeepSeek-V3.2-Speciale, performs&hellip;\n","protected":false},"author":2,"featured_media":224348,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[31],"tags":[26434,345,437,434,435,436,2943,122967,438,146,85,46,11066],"class_list":{"0":"post-224347","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-arts-and-design","8":"tag-agents","9":"tag-ai","10":"tag-arts","11":"tag-arts-and-design","12":"tag-artsanddesign","13":"tag-artsdesign","14":"tag-deepseek","15":"tag-deepseek-v32","16":"tag-design","17":"tag-entertainment","18":"tag-il","19":"tag-israel","20":"tag-ml-data-engineering"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/posts\/224347","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/comments?post=224347"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/posts\/224347\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/media\/224348"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/media?parent=224347"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/categories?post=224347"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/il\/wp-json\/wp\/v2\/tags?post=224347"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}