{"id":156648,"date":"2025-11-24T07:18:13","date_gmt":"2025-11-24T07:18:13","guid":{"rendered":"https:\/\/www.newsbeep.com\/ie\/156648\/"},"modified":"2025-11-24T07:18:13","modified_gmt":"2025-11-24T07:18:13","slug":"latest-open-artifacts-16-whos-building-models-in-the-u-s-chinas-model-release-playbook-and-a-resurgence-of-truly-open-models","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ie\/156648\/","title":{"rendered":"Latest open artifacts (#16):\u00a0 Who&#8217;s building models in the U.S., China&#8217;s model release playbook, and a resurgence of truly open models"},"content":{"rendered":"<p>This holiday season, remember you can give the <a href=\"https:\/\/www.interconnects.ai\/subscribe?gift=true\" rel=\"nofollow noopener\" target=\"_blank\">gift of Interconnects<\/a> (or reimburse your subscription with your company\u2019s learning budget)!<\/p>\n<p>Before we get to the coverage of many truly open models (Stanford\u2019s Marin, Gaperon, Nathan\u2019s Olmo3, etc.), many OCR models, and some frontier models from China\u2019s AI Tigers, we wanted to share a simple list of the AI labs releasing serious open models in the U.S. This list is easily compiled from the backlog of these posts, but having it all in one place is very helpful with the surge of interest in both Chinese and American open model ecosystems.<\/p>\n<p data-attrs=\"{&quot;url&quot;:&quot;https:\/\/www.interconnects.ai\/p\/latest-open-artifacts-16-whos-building?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}\" data-component-name=\"ButtonCreateButton\" class=\"button-wrapper\"><a href=\"https:\/\/www.interconnects.ai\/p\/latest-open-artifacts-16-whos-building?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share\" rel=\"nofollow noopener\" class=\"button primary button-wrapper\" target=\"_blank\">Share<\/a><\/p>\n<p>The U.S. has a comparable number of labs releasing high quality models as China (which is ~20 labs), but many American labs are releasing smaller models with more restrictive licenses, resulting in a far more muted impact. This list includes each notable recent model from them. All of these are pretrained by U.S. companies. To start, the clear players, listed alphabetically:<\/p>\n<p>Ai2 \u2014 Olmo: Open-source leader, to date fully open, smaller dense models mostly. E.g. <a href=\"https:\/\/huggingface.co\/allenai\/Olmo-3-32B-Think\" rel=\"nofollow noopener\" target=\"_blank\">Olmo 3 32B Think<\/a>, the best fully open reasoning model (and best fully open LM ever made). Always developer friendly Apache 2.0 licenses.<\/p>\n<p>Arcee \/ Datology \/ Prime Intellect \u2014 Trinity: Newer startup fully committed to this. Range of MoE models, Apache 2.0 coming soon.<\/p>\n<p>Google \u2014 Gemma: One of the most consistent players. To date models have been smaller and dense as well. E.g. <a href=\"https:\/\/huggingface.co\/google\/vaultgemma-1b\" rel=\"nofollow noopener\" target=\"_blank\">VaultGemma 1B<\/a>. Often solid custom licenses, sometimes minor downstream use restrctions.<\/p>\n<p>HuggingFace \u2014 SmolLM: Tiny models, fully open, great community. E.g. <a href=\"https:\/\/huggingface.co\/HuggingFaceTB\/SmolLM3-3B\" rel=\"nofollow noopener\" target=\"_blank\">SmolLM3-3B<\/a>.<\/p>\n<p>IBM \u2014 Granite: Small to medium sized models, underrated, strong and consistent releases. E.g. <a href=\"https:\/\/huggingface.co\/ibm-granite\/granite-4.0-h-1b\" rel=\"nofollow noopener\" target=\"_blank\">Granite-4.0-h-1b<\/a> with a new hybrid-attention architecture. Developer friendly Apache 2.0 licenses.<\/p>\n<p>Liquid AI \u2014 Liquid Foundation Models: Hybrid architecture, solid small models. .g. <a href=\"https:\/\/huggingface.co\/LiquidAI\/LFM2-VL-3B\" rel=\"nofollow noopener\" target=\"_blank\">LFM2-VL-3B<\/a>, their first vision-language models. Very similar licenses to Apache 2.0, but with restrictions on companies making $10M per year or more.<\/p>\n<p>Microsoft \u2014 Phi (+ others): Solid models, can go under the radar, but consistent. E.g. <a href=\"https:\/\/huggingface.co\/microsoft\/Phi-4-mini-flash-reasoning\" rel=\"nofollow noopener\" target=\"_blank\">Phi-4-mini-flash-reasoning<\/a>, a small reasoning model. Often fairly permissive licenses too, e.g. MIT.<\/p>\n<p>Moondream \u2014 Moondream Models: Solid vision models, originally on device, but dominant and consistent in their niche. E.g. <a href=\"https:\/\/huggingface.co\/moondream\/moondream3-preview\" rel=\"nofollow noopener\" target=\"_blank\">Moondream3<\/a>. Solid licenses and very engaged in community feedback, e.g. recently made it clear synthetic data use is okay.<\/p>\n<p>Nvidia \u2014 Nemotron: Arguably the open leader in the U.S. after Llama 4. E.g. <a href=\"https:\/\/huggingface.co\/nvidia\/NVIDIA-Nemotron-Nano-9B-v2\" rel=\"nofollow noopener\" target=\"_blank\">Nemotron Nano 9B v2<\/a>, a 9B model matching or surpassing Chinese models in that size range. Increasingly open licenses recently, including more open data releases.<\/p>\n<p>OpenAI \u2014 GPT OSS: A new entrant but incredibly important they\u2019re involved. E.g. <a href=\"https:\/\/huggingface.co\/openai\/gpt-oss-120b\" rel=\"nofollow noopener\" target=\"_blank\">gpt-oss-120b<\/a>, OpenAI\u2019s first open weights language model since GPT-2. Good Apache 2.0 licenses, we hope they continue releasing more models.<\/p>\n<p>Reflection \u2014 TBD: but if you can convince someone you\u2019re worth $2B to do this, I believe you.<\/p>\n<p>ServiceNow \u2014 Apriel: Solid reasoning models and other contributions. E.g. <a href=\"https:\/\/huggingface.co\/ServiceNow-AI\/Apriel-H1-15b-Thinker-SFT\" rel=\"nofollow noopener\" target=\"_blank\">Apriel-H1-15b-Thinker<\/a>. <\/p>\n<p>Stanford University \u2014 Marin Community Models: A new entrant in fully-open models like Ai2, scaling up after releasing their first <a href=\"https:\/\/huggingface.co\/marin-community\/marin-32b-base\" rel=\"nofollow noopener\" target=\"_blank\">30B base model<\/a>. We hope they have strong post-training soon!<\/p>\n<p>Of the above, we\u2019re watching Ai2, Nvidia, Arcee, and Reflection the closest, as the players with the most mind-share and momentum on the ground.<\/p>\n<p>Unclear: Companies making fewer contributions currently but have in the past.<\/p>\n<p>Meta \u2014 Llama: The original, but priorities are changing. Crickets since the <a href=\"https:\/\/www.interconnects.ai\/p\/llama-4\" rel=\"nofollow noopener\" target=\"_blank\">Llama 4 fiasco<\/a>.<\/p>\n<p>Reka \u2014 Flash: A few solid models, no updates in a bit.<\/p>\n<p>xAI \u2014 Grok: Many promises of releasing past models, yet to be useful to the community.<\/p>\n<p>This is a list of groups making solid language models in the U.S. Other labs that would easily be included in a more \u201cWestern\u201d ecosystem list would include the likes of Cohere (with solid models, though normally non-commercial licenses), Mistral (with smaller models, normally Apache 2.0), and AI21 labs. There are other types, such as multimodal generation models and biology-focused models that are massive breakthroughs but not listed.<\/p>\n<p>Download this U.S. model maker list as a PDF below:<\/p>\n<p>In case you weren\u2019t aware: Paid subscribers get access to the <a href=\"https:\/\/www.interconnects.ai\/p\/discord\" rel=\"nofollow noopener\" target=\"_blank\">members-only Discord<\/a>, which not only hosts an active community, but also bots covering even more releases in real-time. It also gives a glimpse into the current work and experiments of many labs who (briefly) forget to set their models to private. An example is below. Sometimes we say this is the best perk you get for upgrading to paid \u2014 consider it!<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!UCbt!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07e4f5f0-0e91-479f-b758-8ca22a1c69f2_1884x1134.png\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/ie\/wp-content\/uploads\/2025\/11\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/07e4f5f0-0e91-479f-b758-8ca22a1c69f2_1884.jpeg\" width=\"1456\" height=\"876\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/07e4f5f0-0e91-479f-b758-8ca22a1c69f2_1884x1134.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:876,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:234307,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https:\/\/www.interconnects.ai\/i\/179633798?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07e4f5f0-0e91-479f-b758-8ca22a1c69f2_1884x1134.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   loading=\"lazy\" class=\"sizing-normal\"\/><\/a><\/p>\n<p><a href=\"https:\/\/huggingface.co\/MiniMaxAI\/MiniMax-M2\" rel=\"nofollow noopener\" target=\"_blank\">MiniMax-M2<\/a> by <a href=\"https:\/\/huggingface.co\/MiniMaxAI\" rel=\"nofollow noopener\" target=\"_blank\">MiniMaxAI<\/a>: Probably one of the surprise releases this month: Minimax, whose previous models were usually behind the claimed performance, really took a leap forward with M2, putting them squarely into the spotlight. They also speedrun the (Chinese) model release playbook, something we <a href=\"https:\/\/www.interconnects.ai\/p\/latest-open-artifacts-12-qwen3-235b-a22b-instruct-2507\" rel=\"nofollow noopener\" target=\"_blank\">observed<\/a> in the <a href=\"https:\/\/www.interconnects.ai\/p\/artifacts-7\" rel=\"nofollow noopener\" target=\"_blank\">past<\/a> and was perfected by the likes of Alibaba (Qwen), Moonshot (Kimi) and Zhipu (GLM):<\/p>\n<p>Build a social media presence, mainly on Twitter. This increasingly means that researchers are also active aside from the corporate \/ brand accounts.<\/p>\n<p>Release a new model with (Western) launch partners and ecosystem support on day zero, from vLLM to OpenRouter and tools like Cline. To really get it off the ground, offer free access to the API for a limited time.<\/p>\n<p>Offer a coding subscription which is compatible with Claude Code (or <a href=\"https:\/\/github.com\/QwenLM\/qwen-code\" rel=\"nofollow noopener\" target=\"_blank\">fork a CLI<\/a>) while undercutting their pricing.<\/p>\n<p>Develop your own <a href=\"https:\/\/github.com\/MoonshotAI\/kimi-cli\" rel=\"nofollow noopener\" target=\"_blank\">tooling<\/a> and train the next model to work even better with it.<\/p>\n<p>This strategy, of course, is working: Zhipu <a href=\"https:\/\/www.scmp.com\/tech\/big-tech\/article\/3331324\/zhipu-ai-sees-tenfold-surge-overseas-users-chinese-ai-gains-traction\" rel=\"nofollow noopener\" target=\"_blank\">reportedly<\/a> has over 100K international API users and 3M chatbot users.<\/p>\n<p><a href=\"https:\/\/huggingface.co\/moonshotai\/Kimi-K2-Thinking\" rel=\"nofollow noopener\" target=\"_blank\">Kimi-K2-Thinking<\/a> by <a href=\"https:\/\/huggingface.co\/moonshotai\" rel=\"nofollow noopener\" target=\"_blank\">moonshotai<\/a> (more coverage <a href=\"https:\/\/www.interconnects.ai\/p\/kimi-k2-thinking-what-it-means\" rel=\"nofollow noopener\" target=\"_blank\">here<\/a>): The best open model, competitive with some of the best closed models. However, independent evaluation is a problem as third-party API providers struggle to implement the model correctly, something which we have seen, for example, with the release of GPT-OSS. As an example, running the agentic \u201cVending-Bench\u201d from Andon Labs with a third party provider as opposed to the official API makes a huge difference:<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!p80m!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13b68d2f-e949-498c-8579-b0142cc120d3_1054x1362.png\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/ie\/wp-content\/uploads\/2025\/11\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/13b68d2f-e949-498c-8579-b0142cc120d3_1054.jpeg\" width=\"462\" height=\"597.0056925996205\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/13b68d2f-e949-498c-8579-b0142cc120d3_1054x1362.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1362,&quot;width&quot;:1054,&quot;resizeWidth&quot;:462,&quot;bytes&quot;:822229,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https:\/\/www.interconnects.ai\/i\/179633798?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13b68d2f-e949-498c-8579-b0142cc120d3_1054x1362.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   loading=\"lazy\" class=\"sizing-normal\"\/><\/a><\/p>\n<p>This is a huge problem plaguing open models. Moonshot also documents the tool calling accuracy <a href=\"https:\/\/github.com\/MoonshotAI\/K2-Vendor-Verifier\/\" rel=\"nofollow noopener\" target=\"_blank\">in a repo<\/a>, where a lot of providers perform sub-par, including vLLM with a schema accuracy &lt;90%.<\/p>\n<p><a href=\"https:\/\/huggingface.co\/Qwen\/Qwen3-VL-32B-Instruct\" rel=\"nofollow noopener\" target=\"_blank\">Qwen3-VL-32B-Instruct<\/a> by <a href=\"https:\/\/huggingface.co\/Qwen\" rel=\"nofollow noopener\" target=\"_blank\">Qwen<\/a>: The 2B and 32B version of Qwen3 also get an update. Similar to the 8B version covered in <a href=\"https:\/\/www.interconnects.ai\/p\/latest-open-models-15-its-qwens-world\" rel=\"nofollow noopener\" target=\"_blank\">last month\u2019s artifact<\/a>, the 32B vision model has better text benchmark scores than the initial release of the 32B text-only model.<\/p>\n<p><a href=\"https:\/\/huggingface.co\/allenai\/Olmo-3-32B-Think\" rel=\"nofollow noopener\" target=\"_blank\">Olmo-3-32B-Think<\/a> by <a href=\"https:\/\/huggingface.co\/allenai\" rel=\"nofollow noopener\" target=\"_blank\">allenai<\/a>: A series of truly open models, covering everything from data to all models from the model flow. In case you somehow missed it, check out the coverage of the model:<\/p>\n<p><a native=\"true\" href=\"https:\/\/www.interconnects.ai\/p\/olmo-3-americas-truly-open-reasoning?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web\" rel=\"nofollow noopener\" class=\"embedded-post\" target=\"_blank\"><\/p>\n<p>Olmo 3: America\u2019s truly open reasoning models<\/p>\n<p>We present Olmo 3, our next family of fully open, leading language models\u2026<\/p>\n<p>Read more<\/p>\n<p>3 days ago \u00b7 87 likes \u00b7 30 comments \u00b7 Nathan Lambert<\/p>\n<p><\/a><\/p>\n<p><a href=\"https:\/\/huggingface.co\/ibm-granite\/granite-4.0-h-1b\" rel=\"nofollow noopener\" target=\"_blank\">granite-4.0-h-1b<\/a> by <a href=\"https:\/\/huggingface.co\/ibm-granite\" rel=\"nofollow noopener\" target=\"_blank\">ibm-granite<\/a>: Of course, IBM can\u2019t stop training tiny models and to add them to their model families. The new Granite models feature hybrid attention and MoE architecture first time.<\/p>\n<p><a href=\"https:\/\/huggingface.co\/marin-community\/marin-32b-base\" rel=\"nofollow noopener\" target=\"_blank\">marin-32b-base<\/a> by <a href=\"https:\/\/huggingface.co\/marin-community\" rel=\"nofollow noopener\" target=\"_blank\">marin-community<\/a>: A truly open model by Percy Liang\u2019s <a href=\"https:\/\/marin.community\/\" rel=\"nofollow noopener\" target=\"_blank\">Marin<\/a>.<\/p>\n<p><a href=\"https:\/\/huggingface.co\/cerebras\/GLM-4.6-REAP-218B-A32B-FP8\" rel=\"nofollow noopener\" target=\"_blank\">GLM-4.6-REAP-218B-A32B-FP8<\/a> by <a href=\"https:\/\/huggingface.co\/cerebras\" rel=\"nofollow noopener\" target=\"_blank\">cerebras<\/a>: While we wait for GLM-4.6-Air (and <a href=\"https:\/\/www.chinatalk.media\/p\/the-zai-playbook\" rel=\"nofollow noopener\" target=\"_blank\">mini<\/a>, as teased in this Interview from Chinatalk\u2019s Jordan Schneider and Nathan with Zixuan Li from Zhipu), you can use pruned versions of the GLM-4.6 model.<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!lGUL!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77ba018-6061-44c5-96ac-3489c95bb14c_1266x1094.png\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/ie\/wp-content\/uploads\/2025\/11\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/a77ba018-6061-44c5-96ac-3489c95bb14c_1266.png\" width=\"1266\" height=\"1094\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/a77ba018-6061-44c5-96ac-3489c95bb14c_1266x1094.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1094,&quot;width&quot;:1266,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:138676,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https:\/\/www.interconnects.ai\/i\/179633798?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77ba018-6061-44c5-96ac-3489c95bb14c_1266x1094.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   loading=\"lazy\" class=\"sizing-normal\"\/><\/a><\/p>\n","protected":false},"excerpt":{"rendered":"This holiday season, remember you can give the gift of Interconnects (or reimburse your subscription with your company\u2019s&hellip;\n","protected":false},"author":2,"featured_media":156649,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[220,218,219,61,60,80],"class_list":{"0":"post-156648","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ie","12":"tag-ireland","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/156648","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/comments?post=156648"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/156648\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media\/156649"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media?parent=156648"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/categories?post=156648"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/tags?post=156648"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}