{"id":406944,"date":"2026-01-11T21:08:23","date_gmt":"2026-01-11T21:08:23","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/406944\/"},"modified":"2026-01-11T21:08:23","modified_gmt":"2026-01-11T21:08:23","slug":"distinct-ai-models-seem-to-converge-on-how-they-encode-reality","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/406944\/","title":{"rendered":"Distinct AI Models Seem To Converge On How They Encode Reality"},"content":{"rendered":"<p>Read a story about dogs, and you may remember it the next time you see one bounding through a park. That\u2019s only possible because you have a unified concept of \u201cdog\u201d that isn\u2019t tied to words or images alone. Bulldog or border collie, barking or getting its belly rubbed, a dog can be many things while still remaining a dog.<\/p>\n<p>Artificial intelligence systems aren\u2019t always so lucky. These systems learn by ingesting <a href=\"https:\/\/www.quantamagazine.org\/how-can-ai-id-a-cat-an-illustrated-guide-20250430\/\" rel=\"nofollow noopener\" target=\"_blank\">vast troves of data<\/a> in a process called training. Often, that data is all of the same type \u2014 text for language models, images for computer vision systems, and more exotic kinds of data for systems designed to predict the <a href=\"https:\/\/www.quantamagazine.org\/ai-model-links-smell-molecules-with-metabolic-processes-20221010\/\" rel=\"nofollow noopener\" target=\"_blank\">odor of molecules<\/a> or the <a href=\"https:\/\/www.quantamagazine.org\/how-ai-revolutionized-protein-science-but-didnt-end-it-20240626\/\" rel=\"nofollow noopener\" target=\"_blank\">structure of proteins<\/a>. So to what extent do language models and vision models have a shared understanding of dogs?<\/p>\n<p>Researchers investigate such questions by peering inside AI systems and studying how they represent scenes and sentences. A <a href=\"http:\/\/arxiv.org\/abs\/2310.13018\" rel=\"nofollow noopener\" target=\"_blank\">growing body of research<\/a> has found that different AI models can develop similar representations, even if they\u2019re trained using different datasets or entirely different data types. What\u2019s more, a few studies have suggested that those representations are growing more similar as models grow more capable. In a <a href=\"https:\/\/arxiv.org\/abs\/2405.07987\" rel=\"nofollow noopener\" target=\"_blank\">2024 paper<\/a>, four AI researchers at the Massachusetts Institute of Technology argued that these hints of convergence are no fluke. Their idea, dubbed the Platonic representation hypothesis, has inspired a lively debate among researchers and a <a href=\"http:\/\/arxiv.org\/abs\/2502.16282\" rel=\"nofollow noopener\" target=\"_blank\">slew<\/a> <a href=\"https:\/\/arxiv.org\/abs\/2503.05283\" rel=\"nofollow noopener\" target=\"_blank\">of<\/a> <a href=\"https:\/\/arxiv.org\/abs\/2512.03750\" rel=\"nofollow noopener\" target=\"_blank\">follow-up<\/a> <a href=\"http:\/\/arxiv.org\/abs\/2511.02767\" rel=\"nofollow noopener\" target=\"_blank\">work<\/a>.<\/p>\n<p>The team\u2019s hypothesis gets its name from a 2,400-year-old allegory by the Greek philosopher Plato. In it, prisoners trapped inside a cave perceive the world only through shadows cast by outside objects. Plato maintained that we\u2019re all like those unfortunate prisoners. The objects we encounter in everyday life, in his view, are pale shadows of ideal \u201cforms\u201d that reside in some transcendent realm beyond the reach of the senses.<\/p>\n<p>        <img loading=\"lazy\" width=\"1504\" height=\"2264\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2026\/01\/Plato-bySilanion-at-the-Capitoline-Museums.png\" class=\"block fit-x fill-h fill-v is-loaded mxa large-print-img vertical\" alt=\"A bust of Plato.\" decoding=\"async\"  \/>    <\/p>\n<p>An allegory by the Greek philosopher Plato inspired a recent paper about AI systems.<\/p>\n<p>Silanion at the Capitoline Museums<\/p>\n<p>The Platonic representation hypothesis is less abstract. In this version of the metaphor, what\u2019s outside the cave is the real world, and it casts machine-readable shadows in the form of streams of data. AI models are the prisoners. The MIT team\u2019s claim is that very different models, exposed only to the data streams, are beginning to converge on a shared \u201cPlatonic representation\u201d of the world behind the data.<\/p>\n<p>\u201cWhy do the language model and the vision model align? Because they\u2019re both shadows of the same world,\u201d said <a href=\"https:\/\/web.mit.edu\/phillipi\/\" rel=\"nofollow noopener\" target=\"_blank\">Phillip Isola<\/a>, the senior author of the paper.<\/p>\n<p>Not everyone is convinced. One of the main points of contention involves which representations to focus on. You can\u2019t inspect a language model\u2019s internal representation of every conceivable sentence, or a vision model\u2019s representation of every image. So how do you decide which ones are, well, representative? Where do you look for the representations, and how do you compare them across very different models? It\u2019s unlikely that researchers will reach a consensus on the Platonic representation hypothesis anytime soon, but that doesn\u2019t bother Isola.<\/p>\n<p>\u201cHalf the community says this is obvious, and the other half says this is obviously wrong,\u201d he said. \u201cWe were happy with that response.\u201d<\/p>\n<p>The Company Being Kept<\/p>\n<p>If AI researchers don\u2019t agree on Plato, they might find more common ground with his predecessor Pythagoras, whose philosophy supposedly started from the premise \u201cAll is number.\u201d That\u2019s an apt description of the neural networks that power AI models. Their representations of words or pictures are just long lists of numbers, each indicating the degree of activation of a specific artificial neuron.<\/p>\n<p>To simplify the math, researchers typically focus on a single layer of a neural network in isolation, which is akin to taking a snapshot of brain activity in a specific region at a specific moment in time. They write down the neuron activations in this layer as a geometric object called a vector \u2014 an arrow that points in a particular direction in an abstract space. Modern AI models have many thousands of neurons in each layer, so their representations are high-dimensional vectors that are impossible to visualize directly. But vectors make it easy to compare a network\u2019s representations: Two representations are similar if the corresponding vectors point in similar directions.<\/p>\n<p>Within a single AI model, similar inputs tend to have similar representations. In a language model, for instance, the vector representing the word \u201cdog\u201d will be relatively close to vectors representing \u201cpet,\u201d \u201cbark,\u201d and \u201cfurry,\u201d and farther from \u201cPlatonic\u201d and \u201cmolasses.\u201d It\u2019s a <a href=\"https:\/\/www.quantamagazine.org\/how-embeddings-encode-what-words-mean-sort-of-20240918\/\" rel=\"nofollow noopener\" target=\"_blank\">precise mathematical realization<\/a> of an idea memorably expressed more than 60 years ago by the British linguist John Rupert Firth: \u201cYou shall know a word by the company it keeps.\u201d<\/p>\n<p>What about representations in different models? It doesn\u2019t make sense to directly compare activation vectors from separate networks, but researchers have devised indirect ways to assess representational similarity. One popular approach is to embrace the lesson of Firth\u2019s pithy quote and measure whether two models\u2019 representations of an input keep the same company.<\/p>\n<p>Imagine that you want to compare how two language models represent words for animals. First, you\u2019ll compile a list of words \u2014 dog, cat, wolf, jellyfish, and so on. You\u2019ll then feed these words into both networks and record their representations of each word. In each network, the representations will form a cluster of vectors. You can then ask: How similar are the overall shapes of the two clusters?<\/p>\n<p>\u201cIt can kind of be described as measuring the similarity of similarities,\u201d said <a href=\"https:\/\/ilia10000.github.io\/\" rel=\"nofollow noopener\" target=\"_blank\">Ilia Sucholutsky<\/a>, an AI researcher at New York University.<\/p>\n","protected":false},"excerpt":{"rendered":"Read a story about dogs, and you may remember it the next time you see one bounding through&hellip;\n","protected":false},"author":2,"featured_media":406945,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-406944","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/406944","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=406944"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/406944\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/406945"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=406944"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=406944"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=406944"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}