{"id":256017,"date":"2025-11-11T02:28:15","date_gmt":"2025-11-11T02:28:15","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/256017\/"},"modified":"2025-11-11T02:28:15","modified_gmt":"2025-11-11T02:28:15","slug":"ai-may-blunt-our-thinking-skills-heres-what-you-can-do-about-it","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/256017\/","title":{"rendered":"AI may blunt our thinking skills \u2013 here\u2019s what you can do about it"},"content":{"rendered":"<p><img decoding=\"async\" class=\"Image\" alt=\"New Scientist. Science news and long reads from expert journalists, covering developments in science, technology, health and the environment on the website and the magazine.\" width=\"1350\" height=\"900\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2025\/11\/ai-stupidity-faster.gif\"   loading=\"eager\" fetchpriority=\"high\" data-image-context=\"Article\" data-image-id=\"2503715\" data-caption=\"\" data-credit=\"\"\/><\/p>\n<p>Socrates wasn\u2019t the greatest fan of the written word. Famous for leaving no texts to posterity, the great philosopher is said to have believed that a reliance on writing <a href=\"https:\/\/blogs.ubc.ca\/etec540sept13\/2013\/09\/29\/socrates-writing-vs-memory\/\" rel=\"nofollow noopener\" target=\"_blank\">destroys the memory and weakens the mind<\/a>.<\/p>\n<p>Some 2400 years later, Socrates\u2019s fears seem misplaced \u2013 particularly in light of evidence that <a href=\"https:\/\/www.newscientist.com\/article\/2414241-writing-things-down-may-help-you-remember-information-more-than-typing\/\" rel=\"nofollow noopener\" target=\"_blank\">writing things down improves memory formation<\/a>. But his broader mistrust of cognitive technologies lives on. A growing number of psychologists, neuroscientists and philosophers worry that ChatGPT and similar generative AI tools will chip away at our powers of information recall and blunt our capacity for clear reasoning.<\/p>\n<p>What\u2019s more, while Socrates relied on clever rhetoric to make his argument, these researchers are grounding theirs in empirical data. Their studies have uncovered evidence that even trained professionals disengage their critical thinking skills when using generative AI, and revealed that an over-reliance on these AI tools during the learning process reduces brain connectivity and renders information less memorable. Little wonder, then, that when I asked Google\u2019s Gemini chatbot whether AI tools are turning our brains to jelly and our memories to sieves, it admitted they might be. At least, I think it did: I can\u2019t quite remember now.<\/p>\n<p>But all is not lost. Many researchers suspect we can flip the narrative, turning generative AI into a tool that improves our cognitive performance and augments our intelligence. \u201cAI is not necessarily making us stupid, but we may be interacting with it stupidly,\u201d says <a href=\"https:\/\/www.stonybrook.edu\/commcms\/psychology\/faculty\/_faculty_profiles\/lrichmond\" rel=\"nofollow noopener\" target=\"_blank\">Lauren Richmond<\/a> at Stony Brook University, New York. So, where are we going wrong with generative AI tools? And how can we change our habits to make better use of the technology?<\/p>\n<p>The generative AI age<\/p>\n<p>In recent years, generative AI has become deeply embedded in our lives. <a href=\"https:\/\/www.apaservices.org\/practice\/news\/artificial-intelligence-psychologists-work\" rel=\"nofollow noopener\" target=\"_blank\">Therapists use it to look for patterns<\/a> in their notes. Students <a href=\"https:\/\/www.newscientist.com\/article\/2354663-should-schools-ban-chatgpt-or-embrace-the-technology-instead\/\" rel=\"nofollow noopener\" target=\"_blank\">rely on it for essay writing<\/a>. It has even been welcomed by some media organisations, which may be why financial news website Business Insider reportedly <a href=\"https:\/\/www.theverge.com\/news\/779739\/business-insider-ai-writing-stories\" rel=\"nofollow noopener\" target=\"_blank\">now permits its journalists to use AI<\/a> when drafting stories.<\/p>\n<p>In one sense, all of these AI users are following a millennia-old tradition of \u201ccognitive offloading\u201d \u2013 using a tool or physical action to reduce mental burden. Many of us use this strategy in our daily lives. Every time we write a shopping list instead of memorising which items to buy, we are employing cognitive offloading.<\/p>\n<p>Used in this way, cognitive offloading can help us improve our accuracy and efficiency, while simultaneously freeing up brain space to handle more complex cognitive tasks such as problem-solving, says Richmond. But in a review of the behaviour that Richmond published earlier this year with her Stony Brook colleague <a href=\"https:\/\/www.linkedin.com\/in\/ryan-taylor-95b37b292\/\" rel=\"nofollow noopener\" target=\"_blank\">Ryan Taylor<\/a>, she found <a href=\"https:\/\/www.nature.com\/articles\/s44159-025-00432-2\" rel=\"nofollow noopener\" target=\"_blank\">it has negative effects on our cognition too<\/a>.<\/p>\n<p>\u201cWhen you\u2019ve offloaded something, you almost kind of mentally delete it,\u201d says Richmond. \u201cImagine you make that grocery list, but then you don\u2019t take it with you. You\u2019re actually worse off than if you just planned on remembering the items that you needed to buy at the store.\u201d<\/p>\n<p>Research backs this up. To take one example, a study published in 2018 revealed that when we take photos of objects we see during a visit to a museum, we are <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/abs\/pii\/S002210311730505X\" rel=\"nofollow noopener\" target=\"_blank\">worse at remembering what was on display afterwards<\/a>: we have subconsciously given our phones the task of memorising the objects on show.<\/p>\n<p>This can create a spiral whereby the more we offload, the less we use our brains, which in turn makes us offload even more. \u201cOffloading begets offloading \u2013 it can happen,\u201d says <a href=\"https:\/\/profiles.sussex.ac.uk\/p493-andy-clark\" rel=\"nofollow noopener\" target=\"_blank\">Andy Clark<\/a>, a philosopher at the University of Sussex, UK. In 1998, Clark and his colleague <a href=\"https:\/\/www.newscientist.com\/article\/mg25333710-900-david-chalmers-interview-virtual-reality-is-as-real-as-real-reality\/\" rel=\"nofollow noopener\" target=\"_blank\">David Chalmers<\/a> \u2013 now at New York University \u2013 proposed the <a href=\"https:\/\/www.jstor.org\/stable\/3328150\" rel=\"nofollow noopener\" target=\"_blank\">extended mind thesis<\/a>, which argues that our minds extend into the physical world through objects such as shopping lists and photo albums. Clark doesn\u2019t view that as inherently good or bad \u2013 although he is concerned that as we extend into cyberspace with generative AI and other online services, we are making ourselves vulnerable if those services ever become unavailable because of power cuts or cyberattacks.<\/p>\n<p>Cognitive offloading could also make our memory more vulnerable to manipulation. In a 2019 study, researchers at the University of Waterloo, Canada, presented volunteers with a list of words to memorise and allowed them to type out the words to help remember them. The researchers found that when they <a href=\"https:\/\/pubmed.ncbi.nlm.nih.gov\/31330472\/\" rel=\"nofollow noopener\" target=\"_blank\">surreptitiously added a rogue word to the typed list<\/a>, the volunteers were highly confident that the rogue word had actually been on the list all along.<\/p>\n<p><img decoding=\"async\" class=\"Image\" alt=\"a person's hands holding a shopping list \" width=\"1350\" height=\"901\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2025\/11\/SEI_272156735.jpg\"   loading=\"lazy\" data-image-context=\"Article\" data-image-id=\"2501882\" data-caption=\"We cognitively offload whenever we write a shopping list\" data-credit=\"Mikhail Rudenko\/Alamy\"\/><\/p>\n<p class=\"ArticleImageCaption__Title\">We cognitively offload whenever we write a shopping list<\/p>\n<p class=\"ArticleImageCaption__Credit\">Mikhail Rudenko\/Alamy<\/p>\n<\/p>\n<p>As we have seen, concerns about the harms of cognitive offloading go back at least as far as Socrates. But generative AI has supercharged them. In a study posted online this year, <a href=\"http:\/\/www.shirimelumad.com\/\" rel=\"nofollow noopener\" target=\"_blank\">Shiri Melumad<\/a> and <a href=\"https:\/\/drjyun.github.io\/\" rel=\"nofollow noopener\" target=\"_blank\">Jin Ho Yun<\/a> at the University of Pennsylvania asked 1100 volunteers to write a short essay offering advice on planting a vegetable garden after researching the topic either using a standard web search or ChatGPT. The resulting essays tended to be shorter and contained fewer references to facts if they were written by volunteers who used ChatGPT, which the researchers interpreted as evidence that the <a href=\"https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=5104064\" rel=\"nofollow noopener\" target=\"_blank\">AI tool had made the learning process more passive<\/a> \u2013 and the resulting understanding more superficial. Melumad and Yun argued that this is because the AIs synthesise information for us. In other words, we cognitively offload our opportunity to explore and make discoveries about a subject for ourselves.<\/p>\n<p>Sliding capacities<\/p>\n<p>The latest neuroscience is adding weight to these fears. In experiments detailed in <a href=\"https:\/\/arxiv.org\/abs\/2506.08872\" rel=\"nofollow noopener\" target=\"_blank\">a paper<\/a> pending peer review which was released this summer, <a href=\"https:\/\/www.media.mit.edu\/people\/nkosmyna\/overview\/\" rel=\"nofollow noopener\" target=\"_blank\">Nataliya Kos\u2019myna<\/a> at the Massachusetts Institute of Technology and her colleagues used EEG head caps to measure the brain activity of 54 volunteers as they wrote essays on subjects such as \u201cDoes true loyalty require unconditional support?\u201d and \u201cIs having too many choices a problem?\u201d. Some of the participants wrote their essays using just their own knowledge and experience, those in a second group were allowed to use the Google search engine to explore the essay subject, and a third group could use ChatGPT.<\/p>\n<p>The team discovered that the group using ChatGPT <a href=\"https:\/\/arxiv.org\/abs\/2506.08872\" rel=\"nofollow noopener\" target=\"_blank\">had lower brain connectivity during the task<\/a>, while the group relying simply on their own knowledge had the highest. The browser group, meanwhile, was somewhere in between.<\/p>\n<p>\u201cThere is definitely a danger of getting into the comfort of this tool that can do almost everything. And that can have a cognitive cost,\u201d says Kos\u2019myna.<\/p>\n<p>Critics may argue that a reduction in brain activity needn\u2019t indicate a lack of cognitive involvement in an activity, which Kos\u2019myna accepts. \u201cBut it is also important to look at behavioural measures,\u201d she says. For example, when quizzing the volunteers later, she and her colleagues discovered that the ChatGPT users found it harder to quote their essays, suggesting they hadn\u2019t been as invested in the writing process.<\/p>\n<p>There is also emerging \u2013 if tentative \u2013 evidence of a link between heavy generative AI use and poorer critical thinking. For instance, <a href=\"https:\/\/www.sbs.edu\/team\/dr-michael-gerlich\/\" rel=\"nofollow noopener\" target=\"_blank\">Michael Gerlich<\/a> at the SBS Swiss Business School published a study earlier this year assessing the AI habits and critical thinking skills of 666 people from diverse backgrounds.<\/p>\n<p>Gerlich used structured questionnaires and in-depth interviews to quantify the participants\u2019 critical thinking skills, which revealed that those aged between 17 and 25 had critical thinking scores that were roughly 45 per cent lower than participants who were over 46 years old.<\/p>\n<p><img decoding=\"async\" class=\"Image\" alt=\"tourists photographing the Mona Lisa\" width=\"1350\" height=\"900\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2025\/11\/SEI_272155584.jpg\"   loading=\"lazy\" data-image-context=\"Article\" data-image-id=\"2501884\" data-caption=\"We remember less of what we see when we use our cameras\" data-credit=\"Grzegorz Czapski\/Alamy\"\/><\/p>\n<p class=\"ArticleImageCaption__Title\">We remember less of what we see when we use our cameras<\/p>\n<p class=\"ArticleImageCaption__Credit\">Grzegorz Czapski\/Alamy<\/p>\n<\/p>\n<p>\u201cThese [younger] people also reported that they depend more and more on AI,\u201d says Gerlich: they were between 40 and 45 per cent more likely to say they relied on AI tools than older participants. In combination, Gerlich thinks the two findings hint that <a href=\"https:\/\/www.mdpi.com\/2075-4698\/15\/1\/6#:~:text=Cognitive%20offloading%20refers%20to%20the,involvement,%20potentially%20affecting%20critical%20thinking.\" rel=\"nofollow noopener\" target=\"_blank\">over-reliance on AI reduces critical thinking skills<\/a>.<\/p>\n<p>Others stress that it is too early to draw any firm conclusions, particularly since Gerlich\u2019s study showed correlation rather than causation \u2013 and given that some research suggests <a href=\"https:\/\/eric.ed.gov\/?id=EJ1301306\" rel=\"nofollow noopener\" target=\"_blank\">critical thinking skills are inherently underdeveloped in adolescents<\/a>. \u201cWe don\u2019t have the evidence yet,\u201d says <a href=\"https:\/\/www.linkedin.com\/in\/dr-french\/\" rel=\"nofollow noopener\" target=\"_blank\">Aaron French<\/a> at Kennesaw State University in Georgia.<\/p>\n<p>But other research suggests the link between generative AI tools and critical thinking may be real. In a study published earlier this year by a team at Microsoft and Carnegie Mellon University in Pennsylvania, 319 \u201cknowledge workers\u201d (scientists, software developers, managers and consultants) were asked about their experiences with generative AI. The researchers found that people who expressed higher confidence in the technology <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/lee_2025_ai_critical_thinking_survey.pdf\" rel=\"nofollow noopener\" target=\"_blank\">freely admitted to engaging in less critical thinking while using it<\/a>. This fits with Gerlich\u2019s suspicion that an over-reliance on AI tools instils a degree of \u201ccognitive laziness\u201d in people.<\/p>\n<p>Perhaps most worrying of all is that generative AI tools may even influence the behaviour of people who don\u2019t use the tools heavily. In a study published earlier this year, <a href=\"https:\/\/www.zachary-wojtowicz.com\/\" rel=\"nofollow noopener\" target=\"_blank\">Zachary Wojtowicz<\/a> and <a href=\"https:\/\/www.cmu.edu\/dietrich\/sds\/people\/faculty\/simon-dedeo.html\" rel=\"nofollow noopener\" target=\"_blank\">Simon DeDeo<\/a> \u2013 who were both at Carnegie Mellon University at the time, though Wojtowicz has since moved to MIT \u2013 argued that we have learned to value the effort that goes into certain behaviours, like crafting a thoughtful and sincere apology in order to repair social relationships. If we can\u2019t escape the suspicion that someone has offloaded these cognitively tricky tasks onto an AI \u2013 having the technology draft an apology on their behalf, say \u2013 <a href=\"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/32151\" rel=\"nofollow noopener\" target=\"_blank\">we may be less inclined to believe that they are being genuine<\/a>.<\/p>\n<p>Using tools intelligently<\/p>\n<p>One way to avoid all of these problems is to reset our relationship with generative AI tools, using them in a way that enhances rather than undermines cognitive engagement. That isn\u2019t as easy as it sounds. In a new study, Gerlich found that even volunteers who pride themselves on their critical thinking skills have a tendency to <a href=\"https:\/\/www.mdpi.com\/2306-5729\/10\/11\/172\" rel=\"nofollow noopener\" target=\"_blank\">slip into lazy cognitive habits when using generative AI tools<\/a>. \u201cAs soon as they were using generative AI without guidance, most of them directly offloaded,\u201d says Gerlich.<\/p>\n<p>When there is guidance, however, it is a different story. Supplemental work by Kos\u2019myna and her colleagues provides a good example. They asked the volunteers who had written an essay using only their own knowledge to work on a second version of the same essay, this time using ChatGPT to help them. The EEG data showed that these volunteers maintained high brain connectivity even as they used the AI tool.<\/p>\n<p><img decoding=\"async\" class=\"Image\" alt=\"Post-it notes\" width=\"1350\" height=\"900\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2025\/11\/SEI_272155275.jpg\"   loading=\"lazy\" data-image-context=\"Article\" data-image-id=\"2501883\" data-caption=\"Jotting down notes leaves us vulnerable to memory manipulation\" data-credit=\"Kyle Glenn\/Unsplash\"\/><\/p>\n<p class=\"ArticleImageCaption__Title\">Jotting down notes leaves us vulnerable to memory manipulation<\/p>\n<p class=\"ArticleImageCaption__Credit\">Kyle Glenn\/Unsplash<\/p>\n<\/p>\n<p>Clark argues that this is important. \u201cIf people think about [a given subject] on their own before using AI, it makes a huge difference to the interest, originality and structure of their subsequent essays,\u201d he says.<\/p>\n<p>French sees the benefit in this approach too. In a paper he published last year with his colleague, the late J.P. Shim, he argued that the right way to think about generative AI is as a <a href=\"https:\/\/link.springer.com\/article\/10.1007\/s10796-024-10562-2\" rel=\"nofollow noopener\" target=\"_blank\">tool to enhance your existing understanding of a given subject<\/a>. The wrong way, meanwhile, is to view the tool as a convenient shortcut that replaces the need for you to develop or maintain any understanding.<\/p>\n<p>So what are the secrets to using AI the right way? Clark suggests we should begin by being a bit less trusting: \u201cTreat it like a colleague that sometimes has great ideas, but sometimes is entirely off the rails,\u201d he says. He also believes that the more thinking you do before using a generative AI tool, the better what he dubs your \u201chybrid cognition\u201d will be.<\/p>\n<p>That being said, Clark says there are times when it is \u201csafe\u201d to be a bit cognitively lazy. If you need to bring together a lot of publicly available information, you can probably trust an AI to do that, although you should still double-check its results.<\/p>\n<p>Gerlich agrees there are good ways to use AI. He says it is important to be aware of the \u201canchoring effect\u201d \u2013 a cognitive bias that makes us rely heavily on the first piece of information we get when <a href=\"https:\/\/www.newscientist.com\/article\/mg21228381-800-decision-time-how-subtle-forces-shape-your-choices\/\" rel=\"nofollow noopener\" target=\"_blank\">making decisions<\/a>. \u201cThe information you first receive has a huge impact on your thoughts,\u201d he says. This means that even if you think you are using AI in the right way \u2013 critically evaluating the answers it produces for you \u2013 you are still likely to be guided by what the AI told you in the first place, which can serve as an obstacle to truly original thinking.<\/p>\n<p>But there are strategies you can use to avoid this problem too, says Gerlich. If you are writing an essay about the French Revolution\u2019s negative impacts on society, don\u2019t ask the AI for examples of those negative consequences. \u201cAsk it to tell you facts about the French Revolution and other revolutions. Then look for the negatives and make your own interpretation,\u201d he says. A final stage might involve sharing your interpretation with the AI and asking it to identify any gaps in your understanding, or to suggest what a counter-argument might look like.<\/p>\n<p>This may be easier or harder depending on who you are. To use AI most fruitfully, you should know your strengths and weaknesses. For example, if you are experiencing cognitive decline, then offloading may offer benefits, says Richmond. Personality could also play a role. If you enjoy thinking, it is a good idea to use AI to challenge your understanding of a subject instead of asking it to spoon-feed you facts.<\/p>\n<p>Some of this advice may seem like common sense. But Clark says it is important that as many people as possible are aware of it for a simple reason: if more of us use generative AI in a considered way, we may actually help to keep those tools sharp.<\/p>\n<p>If we expect generative AI to provide us with all the answers, he says, then we will end up producing less original content ourselves. Ultimately, this means that the large language models (LLMs) that power these tools \u2013 which are trained using human-generated data \u2013 will start to decline in capacity. \u201cYou begin to get the danger of what some people call model collapse,\u201d he says: the LLMs are forced into feedback loops where they are trained on their own content, and <a href=\"https:\/\/www.ibm.com\/think\/topics\/model-collapse\" rel=\"nofollow noopener\" target=\"_blank\">their ability to provide creative, high-quality answers deteriorates<\/a>. \u201cWe\u2019ve got a real vested interest in making sure that we continue to write new and interesting things,\u201d says Clark.<\/p>\n<p>In other words, the incorrect use of generative AI might be a two-way street. Emerging research suggests there is some substance to the fears that AI is making us stupid \u2013 but it is also possible that the practice of overusing it is making AI tools stupid, too.<\/p>\n<p class=\"ArticleTopics__Heading\">Topics:<\/p>\n<p><a class=\"ArticleTopics__ListItemLink\" href=\"https:\/\/www.newscientist.com\/article-topic\/psychology\/\" rel=\"nofollow noopener\" target=\"_blank\">psychology<\/a>\/<a class=\"ArticleTopics__ListItemLink\" href=\"https:\/\/www.newscientist.com\/article-topic\/artificial-intelligence\/\" rel=\"nofollow noopener\" target=\"_blank\">artificial intelligence<\/a>                <\/p>\n","protected":false},"excerpt":{"rendered":"Socrates wasn\u2019t the greatest fan of the written word. Famous for leaving no texts to posterity, the great&hellip;\n","protected":false},"author":2,"featured_media":256018,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[554,733,4308,369,86,56,54,55],"class_list":{"0":"post-256017","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-psychology","12":"tag-technology","13":"tag-uk","14":"tag-united-kingdom","15":"tag-unitedkingdom"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/256017","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=256017"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/256017\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/256018"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=256017"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=256017"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=256017"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}