{"id":30237,"date":"2025-07-23T03:06:09","date_gmt":"2025-07-23T03:06:09","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/30237\/"},"modified":"2025-07-23T03:06:09","modified_gmt":"2025-07-23T03:06:09","slug":"why-thinking-time-is-integral-to-ai","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/30237\/","title":{"rendered":"Why Thinking Time is Integral to AI"},"content":{"rendered":"<p>                        <img width=\"300\" height=\"200\" src=\"data:image\/svg+xml,%3Csvg%20xmlns=\" http:=\"\" class=\"alignleft wp-post-image\" alt=\"\" decoding=\"async\" fetchpriority=\"high\" data-lazy- data-lazy- data-lazy-src=\"https:\/\/www.newsbeep.com\/us\/wp-content\/uploads\/2025\/07\/Depositphotos_12761044_S-300x200.jpg\"\/>                <\/p>\n<p>It turns out that giving AI models time to think before returning answers provides better quality answers. Lilian Weng explains why.<\/p>\n<p>Imagine if your AI model could pause to think before it speaks. In a recent <a href=\"https:\/\/lilianweng.github.io\/posts\/2025-05-01-thinking\/\" rel=\"nofollow noopener\" target=\"_blank\">article by Lilian Weng<\/a>, \u201cWhy We Think,\u201d she explores how giving language models the ability to \u201cthink\u201d at test time (through strategies such as test-time compute and chain-of-thought (CoT) prompting) can lead to significantly better performance. Instead of rushing to produce answers in a single pass, these approaches allow models to spend time reasoning, reflecting, and revising. Inspired by Daniel Kahneman\u2019s dual-process theory of cognition, this shift parallels how humans alternate between fast, instinctive responses and slower, more deliberate thought.<\/p>\n<p><a\/>Why Thinking Time Helps<\/p>\n<p>Allowing models to use more computation during inference turns thinking itself into a resource. With CoT prompting, models generate intermediate steps before producing an answer\u2014mirroring how humans solve complex problems. Techniques like beam search, best-of sampling, and process reward models help guide this reasoning process by ranking and selecting the most promising paths.<\/p>\n<p>What\u2019s surprising is that smaller models using test-time strategies can rival much larger ones relying solely on direct decoding. Methods like sequential revision (where a model revises its own response) and reinforcement learning on checkable tasks (like math or code) enable this progress. By optimizing for correctness, these models learn not just to answer but to reason their way to better answers.<\/p>\n<p>See also: <a href=\"https:\/\/www.rtinsights.com\/what-are-neoclouds-and-why-does-ai-need-them\/\" rel=\"nofollow noopener\" target=\"_blank\">What Are Neoclouds and Why Does AI Need Them?<\/a><\/p>\n<p><a\/>New Challenges in Faithful Reasoning<\/p>\n<p>But more reasoning brings more questions. As Wang notes, reasoning traces must be faithful (i.e., truly reflective of the model\u2019s internal logic.) Without safeguards, models can fabricate explanations or even learn to hide reward-hacking behavior from evaluators.<\/p>\n<p>To counter this, researchers are designing CoT monitors, injecting \u201cthinking tokens\u201d to slow down reasoning, and exploring latent-variable training to model hidden thought processes explicitly. Some users employ external tools, such as code interpreters or knowledge bases, to offload certain steps entirely. These strategies aim to ensure not only better performance but also transparency, safety, and trust in how AI systems arrive at their conclusions.<\/p>\n<p>In short, the future of smarter AI may hinge less on how quickly a model answers (something that can produce hallucinations and biased results) and more on how well it thinks.<\/p>\n<p>Weng is the co-founder of Thinking Machines Lab\u00a0 and former VP of AI Safety &amp; robotics, applied research at OpenAI. Read her full breakdown of the logic and theory behind AI reasoning <a href=\"https:\/\/lilianweng.github.io\/posts\/2025-05-01-thinking\/\" rel=\"nofollow noopener\" target=\"_blank\">here<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":"It turns out that giving AI models time to think before returning answers provides better quality answers. Lilian&hellip;\n","protected":false},"author":2,"featured_media":30238,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[46],"tags":[24620,191,74],"class_list":{"0":"post-30237","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-artifical-intelligence","9":"tag-computing","10":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/30237","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=30237"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/30237\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/30238"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=30237"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=30237"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=30237"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}