{"id":213614,"date":"2025-10-14T21:12:10","date_gmt":"2025-10-14T21:12:10","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/213614\/"},"modified":"2025-10-14T21:12:10","modified_gmt":"2025-10-14T21:12:10","slug":"ai-systems-and-humans-see-the-world-differently-and-thats-why-ai-images-look-so-garish","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/213614\/","title":{"rendered":"AI systems and humans \u2018see\u2019 the world differently \u2013 and that\u2019s why AI images look so garish"},"content":{"rendered":"<p>How do computers see the world? It\u2019s not quite the same way humans do.<\/p>\n<p>Recent advances in generative artificial intelligence (AI) make it possible to do more things with computer image processing. You might ask an AI tool to describe an image, for example, or to create an image from a description you provide.<\/p>\n<p>As generative AI tools and services become more embedded in day-to-day life, knowing more about how computer vision compares to human vision is becoming essential.<\/p>\n<p>My latest <a href=\"https:\/\/doi.org\/10.1177\/14703572251358425\" rel=\"nofollow noopener\" target=\"_blank\">research<\/a>, published in Visual Communication, uses AI-generated descriptions and images to get a sense of how AI models \u201csee\u201d \u2013 and discovered a bright, sensational world of generic images quite different from the human visual realm.<\/p>\n<p>            <a href=\"https:\/\/images.theconversation.com\/files\/677748\/original\/file-20250702-62-nh1gn3.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"This image features a pixelated selfie featuring an individual with long brown hair and a fringe. The person has their tongue out and is smiling too. Most of the parts of the image are pixelated with red and yellow squares focusing on certain parts of the\" class=\"lazyload\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/10\/file-20250702-62-nh1gn3.png\"  \/><\/a><\/p>\n<p>              Algorithms see in a very different way to humans.<br \/>\n              <a class=\"source\" href=\"https:\/\/betterimagesofai.org\" rel=\"nofollow noopener\" target=\"_blank\">Elise Racine \/ Better Images of AI \/ Emotion: Joy<\/a>, <a class=\"license\" href=\"http:\/\/creativecommons.org\/licenses\/by\/4.0\/\" rel=\"nofollow noopener\" target=\"_blank\">CC BY<\/a><\/p>\n<p>Comparing human and computer vision<\/p>\n<p>Humans see when light waves enter our eyes through the iris, cornea and lens. Light is converted into electrical signals by a light-sensitive surface called the retina inside the eyeball, and then our <a href=\"https:\/\/www.taylorfrancis.com\/books\/edit\/10.4324\/9781410611581\/handbook-visual-communication-kenneth-smith-keith-kenney-sandra-moriarty-gretchen-barbatsis\" rel=\"nofollow noopener\" target=\"_blank\">brains interpret<\/a> these signals into images we see. <\/p>\n<p>Our vision focuses on key aspects such as colour, shape, movement and depth. Our eyes let us detect changes in the environment and identify potential threats and hazards.<\/p>\n<p>Computers work very differently. They process images by standardising them, inferring the context of an image through metadata (such as time and location information in an image file), and comparing images to other images they have previously learned about. Computers focus on things such as edges, corners or textures present in the image. They also look for patterns and try to classify objects.<\/p>\n<p>            <a href=\"https:\/\/images.theconversation.com\/files\/677750\/original\/file-20250702-56-q9q1pw.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"A screenshot of a CAPTCHA test asking a user to select all images with a bus.\" class=\"lazyload\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/10\/file-20250702-56-q9q1pw.png\"  \/><\/a><\/p>\n<p>              Solving CAPTCHAs helps prove you\u2019re human and also helps computers learn how to \u2018see\u2019.<br \/>\n              CAPTCHA<\/p>\n<p>You\u2019ve likely helped computers learn how to \u201csee\u201d by completing online <a href=\"http:\/\/www.captcha.net\/\" rel=\"nofollow noopener\" target=\"_blank\">CAPTCHA tests<\/a>. <\/p>\n<p>These are typically used to help computers differentiate between humans and bots. But they\u2019re also used to train and improve machine learning algorithms.<\/p>\n<p>So, when you\u2019re asked to \u201cselect all the images with a bus\u201d, you\u2019re helping software learn the difference between different types of vehicles as well as proving you\u2019re human. <\/p>\n<p>Exploring how computers \u2018see\u2019 differently<\/p>\n<p>In my new research, I asked a large language model to describe two visually distinct sets of human-created images. <\/p>\n<p>One set contained hand-drawn illustrations while the other was made up of camera-produced photographs.<\/p>\n<p>            <a href=\"https:\/\/images.theconversation.com\/files\/677753\/original\/file-20250702-62-eqryu5.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"A screenshot of several image thumbnails, some illustrations and some photos.\" class=\"lazyload\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/10\/file-20250702-62-eqryu5.png\"  \/><\/a><\/p>\n<p>              Some of the nuances of algorithmic vision can be uncovered by asking an AI tool to describe images and then visualise those same descriptions.<br \/>\n              T.J. Thomson, Author provided (no reuse)<\/p>\n<p>I fed the descriptions back into an AI tool and asked it to visualise what it had described. I then compared the original human-made images to the computer-generated ones.<\/p>\n<p>The resulting descriptions noted the hand-drawn images were illustrations but didn\u2019t mention the other images as being photographs or having a high level of realism. This suggests AI tools see photorealism as the default visual style, unless specifically prompted otherwise.<\/p>\n<p>Cultural context was largely devoid from the descriptions. The AI tool either couldn\u2019t or wouldn\u2019t infer cultural context by the presence of, for example, Arabic or Hebrew writing in the images. This underscores the dominance of some languages, like English, in AI tools\u2019 training data.<\/p>\n<p>While colour is vital to human vision, it too was largely ignored in the AI tools\u2019 image descriptions. Visual depth and perspective were also largely ignored.<\/p>\n<p>The AI images were more boxy than the hand-drawn illustrations, which used more organic shapes. <\/p>\n<p>            <a href=\"https:\/\/images.theconversation.com\/files\/677755\/original\/file-20250702-62-qt9m69.jpg?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"Two similar but different black and white illustrations of a bookshelf on wheels.\" class=\"lazyload\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/10\/file-20250702-62-qt9m69.jpg\"  \/><\/a><\/p>\n<p>              The AI-generated images were much more boxy than the hand-drawn illustrations, which used more organic shapes and had a different relationship between positive and negative space.<br \/>\n              Left: Medar de la Cruz; right: ChatGPT<\/p>\n<p>The AI images were also much more saturated than the source images: they contained brighter, more vivid colours. This reveals the prevalence of stock photos, which tend to be more \u201ccontrasty\u201d, in AI tools\u2019 <a href=\"https:\/\/www.tandfonline.com\/doi\/full\/10.1080\/21670811.2023.2229883\" rel=\"nofollow noopener\" target=\"_blank\">training data<\/a>. <\/p>\n<p>The AI images were also more sensationalist. A single car in the original image became one of a long column of cars in the AI version. AI seems to exaggerate details not just in text but also in visual form.<\/p>\n<p>            <a href=\"https:\/\/images.theconversation.com\/files\/677756\/original\/file-20250702-56-obobkl.jpg?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"A photo of people with guns driving through a desert and a generated photorealistic image of several cars containing peopl with guns driving through a desert.\" class=\"lazyload\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/10\/file-20250702-56-obobkl.jpg\"  \/><\/a><\/p>\n<p>              The AI-generated images were more sensationalist and contrasty than the human-created photographs.<br \/>\n              Left: Ahmed Zakot; right: ChatGPT<\/p>\n<p>The generic nature of the AI images means they can be used in many contexts and across countries. But the lack of specificity also means <a href=\"https:\/\/apo.org.au\/node\/329601\" rel=\"nofollow noopener\" target=\"_blank\">audiences might perceive them<\/a> as less authentic and engaging.<\/p>\n<p>Deciding when to use human or computer vision<\/p>\n<p>This research supports the notion that humans and computers \u201csee\u201d differently. Knowing when to rely on computer or human vision to describe or create images can be a competitive advantage.<\/p>\n<p>While AI-generated images can be eye-catching, they can also come across as hollow upon closer inspection. This can limit their value.<\/p>\n<p>Images are adept at sparking an emotional reaction and audiences might find human-created images that authentically reflect specific conditions as <a href=\"https:\/\/apo.org.au\/node\/329601\" rel=\"nofollow noopener\" target=\"_blank\">more engaging<\/a> than computer-generated attempts.<\/p>\n<p>However, the capabilities of AI can make it an attractive option for quickly labelling large data sets and helping humans categorise them.<\/p>\n<p>Ultimately, there\u2019s a role for both human and AI vision. Knowing more about the opportunities and limits of each can help keep you safer, more productive, and better equipped to communicate in the digital age.<\/p>\n","protected":false},"excerpt":{"rendered":"How do computers see the world? It\u2019s not quite the same way humans do. Recent advances in generative&hellip;\n","protected":false},"author":2,"featured_media":213615,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-213614","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/213614","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=213614"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/213614\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/213615"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=213614"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=213614"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=213614"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}