{"id":543494,"date":"2026-04-21T21:06:10","date_gmt":"2026-04-21T21:06:10","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/543494\/"},"modified":"2026-04-21T21:06:10","modified_gmt":"2026-04-21T21:06:10","slug":"chatgpts-new-images-2-0-model-is-surprisingly-good-at-generating-text","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/543494\/","title":{"rendered":"ChatGPT&#8217;s new Images 2.0 model is surprisingly good at generating text"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">It used to be easy enough to distinguish between human-made and AI-generated imagery \u2014 just two years ago, you couldn\u2019t use image models to <a href=\"https:\/\/techcrunch.com\/2024\/03\/21\/why-is-ai-so-bad-at-spelling\/\" rel=\"nofollow noopener\" target=\"_blank\">create a menu for a Mexican restaurant<\/a> without inventing new culinary delights like \u201cenchuita,\u201d \u201cchuriros,\u201d \u201cburrto,\u201d and \u201cmargartas.\u201d <\/p>\n<p class=\"wp-block-paragraph\">Now, when I ask the brand new ChatGPT Images 2.0 model for a menu of Mexican food, it creates something that could immediately be used in a restaurant without customers noticing that something\u2019s off. (However, ceviche priced at $13.50 might make me question the quality of the fish.)<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"680\" width=\"453\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2026\/04\/6e744049-12b8-49ba-8d2a-66e7326c0169.png\" alt=\"\" class=\"wp-image-3114706\"  \/>Image Credits:ChatGPT Images 2.0<\/p>\n<p class=\"wp-block-paragraph\">For comparison, here\u2019s the result I got from DALL-E 3 two years ago (at the time, ChatGPT did not generate images):<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"680\" width=\"652\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2026\/04\/Screenshot-2024-03-19-at-11.05.24-AM.webp.png\" alt=\"\" class=\"wp-image-3114711\"  \/>Image Credits:Microsoft Designer (DALL-E 3)<\/p>\n<p class=\"wp-block-paragraph\">AI image generators have <a href=\"https:\/\/techcrunch.com\/2024\/03\/21\/why-is-ai-so-bad-at-spelling\/\" rel=\"nofollow noopener\" target=\"_blank\">historically struggled to spell<\/a> because they generally used diffusion models, which work by reconstructing images from noise.<\/p>\n<p class=\"wp-block-paragraph\">\u201cThe diffusion models [\u2026] are reconstructing a given input,\u201d Asmelash Teka Hadgu, founder and CEO of Lesan AI, <a href=\"https:\/\/techcrunch.com\/2024\/03\/21\/why-is-ai-so-bad-at-spelling\/\" rel=\"nofollow noopener\" target=\"_blank\">told TechCrunch<\/a> in 2024. \u201cWe can assume writings on an image are a very, very tiny part, so the image generator learns the patterns that cover more of these pixels.\u201d<\/p>\n<p class=\"wp-block-paragraph\">Researchers have since explored other mechanisms for image generation, like <a href=\"https:\/\/aws.amazon.com\/what-is\/autoregressive-models\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">autoregressive models<\/a>, which make predictions about what an image should look like and function more like an LLM.<\/p>\n<p class=\"wp-block-paragraph\">Unfortunately, OpenAI declined to answer a question in a press briefing this week about what kind of model is powering ChatGPT Images 2.0.<\/p>\n<p>Techcrunch event<\/p>\n<p>\n\t\t\t\t\t\t\t\t\tSan Francisco, CA<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\t|<br \/>\n\t\t\t\t\t\t\t\t\t\t\t\t\tOctober 13-15, 2026\n\t\t\t\t\t\t\t<\/p>\n<p class=\"wp-block-paragraph\">The company did, however, explain that the new model has \u201cthinking capabilities,\u201d which give it the ability to search the web, make multiple images from one prompt, and double-check its creations \u2014 this allows Images 2.0 to create marketing assets in various sizes, as well as multi-paneled comic strips. <\/p>\n<p class=\"wp-block-paragraph\">OpenAI also says that Images has a stronger understanding of non-Latin text rendering in languages like Japanese, Korean, Hindi, and Bengali. The model\u2019s knowledge cuts off in December 2025, which could impact how accurately it can generate certain prompts involving recent news.<\/p>\n<p class=\"wp-block-paragraph\">\u201cImages 2.0 brings an unprecedented level of specificity and fidelity to image creation. It can not only conceptualize more sophisticated images, but it actually brings that vision to life e\ufb00ectively, able to follow instructions, preserve requested details, and render the fine-grained elements that often break image models: small text, iconography, UI elements, dense compositions, and subtle stylistic constraints, all at up to 2K resolution,\u201d OpenAI said in a press release.<\/p>\n<p class=\"wp-block-paragraph\">These capabilities mean that image generation isn\u2019t as rapid as typing a question to ChatGPT, but generating something complex like a multi-paneled comic still takes just a few minutes. <\/p>\n<p class=\"wp-block-paragraph\">All ChatGPT and Codex users will be able to access Images 2.0 starting Tuesday; paid users will be able to generate more advanced outputs. The company will also make the gpt-image-2 <a href=\"https:\/\/openai.com\/api\/pricing\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">API available<\/a>, with pricing dependent on the quality and resolution of outputs.<\/p>\n<p>When you purchase through links in our articles, <a href=\"https:\/\/techcrunch.com\/techcrunch-affiliate-monetization-standards\/\" rel=\"nofollow noopener\" target=\"_blank\">we may earn a small commission<\/a>. This doesn\u2019t affect our editorial independence.<\/p>\n","protected":false},"excerpt":{"rendered":"It used to be easy enough to distinguish between human-made and AI-generated imagery \u2014 just two years ago,&hellip;\n","protected":false},"author":2,"featured_media":517710,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[554,733,4308,1921,126580,874,86,56,54,55],"class_list":{"0":"post-543494","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-chatgpt","12":"tag-image-generation","13":"tag-openai","14":"tag-technology","15":"tag-uk","16":"tag-united-kingdom","17":"tag-unitedkingdom"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/543494","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=543494"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/543494\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/517710"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=543494"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=543494"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=543494"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}