{"id":368508,"date":"2026-03-27T18:52:13","date_gmt":"2026-03-27T18:52:13","guid":{"rendered":"https:\/\/www.newsbeep.com\/ie\/368508\/"},"modified":"2026-03-27T18:52:13","modified_gmt":"2026-03-27T18:52:13","slug":"what-is-anthropic-thinking-derek-thompson","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ie\/368508\/","title":{"rendered":"What Is Anthropic Thinking? &#8211; Derek Thompson"},"content":{"rendered":"<p>The landscape of AI is not merely filled with news. It is filled with teams. You have the doomers, the accelerationists, the skeptics, the it\u2019s-a-bubble oracles, the anti-bubble counter oracles, and so on. It would be convenient for my sanity\u2014and, perhaps, the sanity of my readers\u2014if I simply joined one team and never removed the jersey. But I don\u2019t think any aforementioned tribe has a monopoly on good arguments. I think the doomers are right about the risk of the technology, and the accelerationists are right about the promise of the technology, and the skeptics are right that the doomers and accelerationists can both overstate their cases.<\/p>\n<p>So I\u2019m trying to make sure that my coverage of AI cuts across silos and allows readers to hear from all of the teams. Last week, I published an interview with the investor and writer Paul Kedrosky on the case for AI being an economic bubble. But if any single data point pierces that narrative, it\u2019s this: Between December 2025 and March 2026, the AI lab Anthropic more than doubled its annual recurring revenue from $9 billion to more than $20 billion. According to several analysts, there is no record of any company growing this fast at this scale \u2026 ever. It\u2019s hard to imagine that artificial intelligence is both a bubble and the home to the industry witnessing the fastest-growing businesses in history.<\/p>\n<p>Today\u2019s conversation is with Anthropic co-founder Jack Clark. I don\u2019t need Jack or anybody at Anthropic to read me a corporate statement about the company\u2019s revenue growth. I can do that, myself. What I really wanted to hear from Jack were his answers to the deepest, thorniest philosophical questions about the meaning of AI. Questions like:<\/p>\n<p>If Anthropic\u2019s executives believes that AI might be as dangerous as nuclear weapons, what right does any private business have to build this sort of thing for profit?<\/p>\n<p>If AI is really so good at making people more productive, why do Americans overall say they disapprove of AI more than just about every other institution and individual in the world?<\/p>\n<p>Does Anthropic really believe that AI will lead to imminent mass unemployment? <\/p>\n<p>Why\u2014as Noah Smith recently <a href=\"https:\/\/www.noahpinion.blog\/p\/ai-has-the-worst-sales-pitch-ive\" rel=\"nofollow noopener\" target=\"_blank\">put it<\/a>\u2014does this industry insist on \u201cour product will make you economically useless, and possibly kill you!\u201d as a marketing strategy?<\/p>\n<p>Why does AI still seem quite inept at coming up with truly original insights?<\/p>\n<p>How does Anthropic use its own autonomous agents to increase productivity within the company?<\/p>\n<p>If other companies learn to use agents effectively, is knowledge work <a href=\"https:\/\/www.instagram.com\/reel\/DSvlXwtDeSM\/\" rel=\"nofollow noopener\" target=\"_blank\">\u201ccooked\u201d<\/a>?<\/p>\n<p>How should we raise our children in an age of AI? <\/p>\n<p>And what values would super-intelligence make even more important than they are today?<\/p>\n<p>Derek Thompson: Anthropic has compared artificial intelligence to nuclear weapons on several occasions. Most recently in January, Dario Amodei, the CEO of Anthropic, said that the Trump administration\u2019s decision to allow advanced NVIDIA chips to be exported to China was \u201ca bit like selling nuclear weapons to North Korea.\u201d<\/p>\n<p>The US does not allow private companies to build nuclear weapons. That is the law. If artificial intelligence is just like nuclear weapons, why should we allow private firms to build it for profit?<\/p>\n<p>Jack Clark: AI is fundamentally like everything. It\u2019s like a factory that produces cars, micro scooters, animals, and nuclear weapons all at the same time. And the main question we\u2019re going to have to deal with as a society is how do you govern those factories and how do you decide what the appropriate uses are of the things that come out? I can\u2019t talk about the specifics of our ongoing discussions with the Department of Defense. I can say that Anthropic was extremely committed to working on national security early because we recognize that AI is going to touch every single part of life, and every single part of life is going to have its own range of incredibly thorny, difficult issues. We\u2019re going to need a much larger societal conversation about how we govern this technology in general, and we will need to reckon with the fact that the technology comes from the private sector and then flows into all of these other sectors. That\u2019s going to be really challenging. It\u2019s something we haven\u2019t encountered before, because previously you didn\u2019t have a technology that could become anything. You had specific technologies built by specific industries for specific purposes, and that was in many ways simpler.<\/p>\n<p>Thompson: I want to hear a robust defense of why this is the private sector\u2019s job. The nuclear analogy is invoked in so many different ways: for export controls, for arguments about government attention, for arguments about existential stakes, for arguments about the need for international cooperation. But one conclusion this analogy very clearly supports is that the private sector should not control this technology. <\/p>\n<p>So why does the analogy apply almost everywhere except here, where private sectors are developing frontier AI for profit while the government attempts to regulate it from the outside?<\/p>\n<p>Clark: We worked for many years with the National Nuclear Security Administration to actually test out how well AI could understand aspects of nuclear technology. And we used that to develop evals and ways of ensuring that we don\u2019t proliferate things into the world that have an understanding of nuclear weapons. That\u2019s almost a very positive example of how you would have the private sector work with government, where some things absolutely should only be the domain of government, like nuclear weapons. The job of a company producing a technology that can take on many different aspects is to work out the areas where it\u2019s inappropriate to deploy that technology, like nuclear weapons, and then work with government to take that capability surface off. I think that describes some of the path we\u2019re going to have to pursue here. And it\u2019s one that most of the industry is going down, including for biological weapons.<\/p>\n<p>Thompson: So you\u2019re saying the right way to think about this is that AI is this is a multifarious factory technology, where you are creating super-powered Excel charts, which is a technology that has no precedent for government regulation, but you\u2019re also creating technology that can be used by the Pentagon or by individuals to have militaristic or dangerous ends. And so the analogy with nuclear weapons is true insofar as it is contained to the parts of your technology that are like nuclear weapons, but you\u2019re also doing a lot of other things that have no analogy in nuclear weapons, like making white collar workers a little bit more productive at their desk jobs. Is that a fair summary?<\/p>\n<p>Clark: There are almost two problems here. One is that you have this factory that can produce anything, so you make sure that what comes out correlates to what we\u2019ve decided society can have available in the free market. Not nuclear weapons. Yes, things that accelerate knowledge workers. And then you have this second question of, given the multifaceted nature of what can be produced, how do you work with government or academia or other parties on the things which you can\u2019t push out to the world in general, but which have value in the rest of the world? An example here is biology. For [technology] that can massively accelerate the development of biological science, you need to work out: what is the path to bringing that technology about? So some of the conversation that society is going to have now is, what are the appropriate ways we want this technology to be used, and how do people decide what to do with different things in this factory, and how to proliferate them, so society gets the benefit?<\/p>\n<p>Thompson: Anthropic CEO Dario Amodei has predicted on several occasions that AI will destroy half of all entry-level white collar positions and spike unemployment to as high as 20%, which would be the highest unemployment rate since the Great Depression. He said this could happen in as soon as five years. Do you agree with that forecast?<\/p>\n<p>Clark: We\u2019re talking about one of the potential things that can happen, and I think it\u2019s worth noting that this is a choice. I don\u2019t agree with this, because I think it\u2019s a choice that we can make. Also, my personal view based on the data I look at is that big changes in employment take a long time to filter through to the economy. And even with the magnitude of what we\u2019re talking about, you might expect it to take longer. <\/p>\n<p>Let\u2019s say there is the potential for massive employment changes. I think that this is accompanied by the fact that AI must also be growing the economy a lot and causing a lot of economic activity. If that is the case, then you would expect more freedom about [designing] policy and [deciding] what we do with this economy. If you end up in a situation where employment is negatively affected by AI in one part of the economy, but loads of money being generated by AI in another part of the economy, you could choose to create jobs, like teaching or nursing, where we have a preference for more people working. You could both increase the number of jobs and do things like cross-sector wage subsidies to improve the wages of those jobs, where today we severely under-compensate them.<\/p>\n<p>Thompson: It is unusual in corporate history for a company to announce that if its product is successful, tens of millions of people will lose their jobs and there\u2019s a non-zero chance that we end the human race entirely. I think the analogy I\u2019ve used before is that Henry Ford would have been within the realm of reality if he said, \u201cIf this Model T thing takes off, hundreds of thousands of Americans are going to die every decade from car accidents.\u201d In fact, that\u2019s happened. But Ford and GM did not talk like that in the 1910s and 1920s. What is the strategy of communicating your technology to the American people as a means by which we might have 20 percent unemployment and a non-zero chance of human catastrophe?<\/p>\n<p>Clark: These are not the outcomes we want or anyone in the industry wants. But I think the industry has also learned from the overly rosy predictions made by many in the technology industry before, about how the only effect they\u2019d have on the world would be unalloyed positivity. And I think the world lost huge amounts of trust in the technology industry because of that, because they saw that it wasn\u2019t only positive. Social media has caused a range of amazing positives in the world and a range of harms, which we\u2019re now dealing with. The ethos here, and why I\u2019m working on a new initiative called the Anthropic Institute, is to share a lot more data about what we see in front of us so that society is better prepared for any of the different changes which could come along.<\/p>\n<p>We also don\u2019t spend enough time talking about all of the really positive changes, which I think are a choice that we can make as a civilization. But it would be negligent of us to not call out that there are ways we as a species could get this technology wrong. If you look at scientists and people that have worked on transformational technologies before in biology or in the early days of nanotechnology, they\u2019ve all talked about this combination of upsides and risks. It\u2019s just that AI as a sector has matured and made a lot more impact on the markets than either of those classes of technology over the same time period, so everything\u2019s accentuated.<\/p>\n<p>Thompson: I hear the argument that you are reacting to the social media experience. But I also look at polling. Last week, NBC News published a national survey on attitudes toward a range of politicians and institutions. AI\u2019s net favorability was negative 20. That\u2019s below every politician surveyed in the poll and it\u2019s below ICE. Why do you think people say they hate artificial intelligence?<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!bmbd!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859a9830-7745-4f86-9586-09c473f01edd_500x272.png\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/ie\/wp-content\/uploads\/2026\/03\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/859a9830-7745-4f86-9586-09c473f01edd_500x.jpeg\" width=\"500\" height=\"272\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/859a9830-7745-4f86-9586-09c473f01edd_500x272.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:272,&quot;width&quot;:500,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75248,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https:\/\/www.derekthompson.org\/i\/192306504?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859a9830-7745-4f86-9586-09c473f01edd_500x272.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   loading=\"lazy\" class=\"sizing-normal\"\/><\/a><script async src=\"\/\/www.instagram.com\/embed.js\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"The landscape of AI is not merely filled with news. It is filled with teams. You have the&hellip;\n","protected":false},"author":2,"featured_media":368509,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[220,218,219,61,60,80],"class_list":{"0":"post-368508","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ie","12":"tag-ireland","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/368508","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/comments?post=368508"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/368508\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media\/368509"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media?parent=368508"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/categories?post=368508"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/tags?post=368508"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}