<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Anthropic on Carles Abarca</title><link>https://carlesabarca.com/tags/anthropic/</link><description>Recent content in Anthropic on Carles Abarca</description><generator>Hugo -- gohugo.io</generator><language>en</language><copyright>© 2026 Carles Abarca</copyright><lastBuildDate>Thu, 09 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://carlesabarca.com/tags/anthropic/index.xml" rel="self" type="application/rss+xml"/><item><title>Claude Mythos: the model Anthropic chose not to release</title><link>https://carlesabarca.com/posts/claude-mythos-unreleased-frontier-model/</link><pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/claude-mythos-unreleased-frontier-model/</guid><description>Anthropic has done something extraordinary: publish technical documentation about its most advanced model while refusing to deploy it broadly. Claude Mythos Preview may mark a turning point in the relationship between capability, security, and frontier model release.</description><content:encoded>&lt;blockquote&gt;&lt;p&gt;“Claude Mythos Preview is a general-purpose, unreleased frontier model.”&lt;br&gt;
— Anthropic, &lt;em&gt;Project Glasswing&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;Anthropic has just made a decision that, until very recently, would have seemed almost unthinkable in the race for frontier models: &lt;strong&gt;publicly present a new-generation model while simultaneously deciding not to make it broadly available to the market&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;This is not a product delay. Nor is it a conventional beta program. What Anthropic has done with &lt;strong&gt;Claude Mythos Preview&lt;/strong&gt; is something else: publish part of the technical documentation, describe extraordinary capabilities—especially in offensive cybersecurity—and restrict access to a very limited circle of defensive actors under a specific initiative: &lt;strong&gt;Project Glasswing&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;The important question is not only what Mythos is. The important question is &lt;strong&gt;what it means that Anthropic has decided not to launch it like a normal model&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;The extraordinary part is not the model. It is the decision.
 &lt;div id="the-extraordinary-part-is-not-the-model-it-is-the-decision" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-extraordinary-part-is-not-the-model-it-is-the-decision" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;In the AI industry, a fairly clear logic had taken hold: if a lab trains a better model, sooner or later it turns it into a product. It may do so gradually, via APIs, waitlists, enterprise agreements, or usage restrictions. But the overall direction was unmistakable: &lt;strong&gt;more capability eventually meant more availability&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;With Mythos, Anthropic introduces a break.&lt;/p&gt;
&lt;p&gt;On the one hand, it presents the model as a new frontier of capability. On the other, it implicitly admits that &lt;strong&gt;this capability crosses a threshold that makes broad deployment irresponsible&lt;/strong&gt;.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“We formed Project Glasswing because of capabilities we’ve observed in a new frontier model trained by Anthropic that we believe could reshape cybersecurity.”&lt;br&gt;
— Anthropic, &lt;em&gt;Project Glasswing&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;That is not routine marketing language. It is a governance signal. Anthropic is saying that, in its judgment, the model is not just better: &lt;strong&gt;it is dangerously better in one specific dimension&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;What Anthropic claims about Claude Mythos Preview
 &lt;div id="what-anthropic-claims-about-claude-mythos-preview" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-anthropic-claims-about-claude-mythos-preview" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The documentation Anthropic has published paints a picture that is difficult to ignore.&lt;/p&gt;
&lt;p&gt;In its Frontier Red Team technical post, the company argues that Mythos Preview:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;identifies and exploits &lt;strong&gt;zero-days&lt;/strong&gt; in real software,&lt;/li&gt;
&lt;li&gt;does so across &lt;strong&gt;every major operating system&lt;/strong&gt; and &lt;strong&gt;every major browser&lt;/strong&gt;,&lt;/li&gt;
&lt;li&gt;produces complex exploits, including multi-vulnerability chains,&lt;/li&gt;
&lt;li&gt;and represents a radical leap beyond previous Claude generations.&lt;/li&gt;
&lt;/ul&gt;
&lt;blockquote&gt;&lt;p&gt;“During our testing, we found that Mythos Preview is capable of identifying and then exploiting zero-day vulnerabilities in every major operating system and every major web browser when directed by a user to do so.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;If this is correct, we are not looking at an incremental improvement. We are looking at a &lt;strong&gt;regime change&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Anthropic goes further. It says internal engineers with no formal security training have asked the model to find a remote vulnerability overnight and woken up the next morning to a complete working exploit.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“Engineers at Anthropic with no formal security training have asked Mythos Preview to find remote code execution vulnerabilities overnight, and woken up the following morning to a complete, working exploit.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;That detail matters. It suggests not only that the model amplifies expert capability. It also suggests that it &lt;strong&gt;dramatically lowers the barrier to entry&lt;/strong&gt; for advanced offensive capability.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;The leap beyond Opus 4.6
 &lt;div id="the-leap-beyond-opus-46" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-leap-beyond-opus-46" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;One of the most striking elements of the technical documentation is the comparison with earlier generations.&lt;/p&gt;
&lt;p&gt;Anthropic notes that, only a month earlier, its read on &lt;strong&gt;Opus 4.6&lt;/strong&gt; was that the model was much better at finding and fixing vulnerabilities than at exploiting them. In other words, it was still strong in defensive cybersecurity, but not especially effective at autonomous offensive work.&lt;/p&gt;
&lt;p&gt;With Mythos, that changes.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“Opus 4.6 generally had a near-0% success rate at autonomous exploit development. But Mythos Preview is in a different league.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;The company cites a benchmark involving Firefox vulnerabilities where Opus 4.6 only managed to convert findings into working exploits a handful of times, while Mythos Preview did so &lt;strong&gt;181 times&lt;/strong&gt;, with register control in &lt;strong&gt;29 additional cases&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;If those numbers hold, we are not talking about “a stronger Claude.” We are talking about &lt;strong&gt;a different order of capability&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;It was not trained “to hack”
 &lt;div id="it-was-not-trained-to-hack" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#it-was-not-trained-to-hack" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;This point is critical.&lt;/p&gt;
&lt;p&gt;Anthropic says it did &lt;strong&gt;not explicitly train Mythos Preview to develop these offensive capabilities&lt;/strong&gt;. According to the company, what we are seeing is an emergent consequence of broader improvements in:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;reasoning,&lt;/li&gt;
&lt;li&gt;autonomy,&lt;/li&gt;
&lt;li&gt;code work,&lt;/li&gt;
&lt;li&gt;and multi-step planning.&lt;/li&gt;
&lt;/ul&gt;
&lt;blockquote&gt;&lt;p&gt;“We did not explicitly train Mythos Preview to have these capabilities. Rather, they emerged as a downstream consequence of general improvements in code, reasoning, and autonomy.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;That sentence deserves to be read carefully, because it points to something bigger than Mythos. It suggests that &lt;strong&gt;as generalist models improve at useful code work and agentic behavior, offensive capability stops being a separate specialty&lt;/strong&gt;. It appears as a natural side effect of general progress.&lt;/p&gt;
&lt;p&gt;That makes governance much harder. It is no longer enough to avoid training “a model for cyberattack.” The real issue is that &lt;strong&gt;a sufficiently capable general model can become a first-rate offensive tool even if that was never the explicit objective of training&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;So why not release it?
 &lt;div id="so-why-not-release-it" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#so-why-not-release-it" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Anthropic frames the answer in terms of a &lt;strong&gt;dangerous transition window&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Its thesis is that, in the long run, tools like this may benefit defenders more than attackers. But in the short run there is an obvious risk: offensive capability may diffuse faster than defensive capability can absorb it.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“In the short term, this could be attackers, if frontier labs aren’t careful about how they release these models.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;That is why there is no broad release. Instead, Anthropic created &lt;strong&gt;Project Glasswing&lt;/strong&gt;, an initiative involving partners such as AWS, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks, along with dozens of additional organizations.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“By releasing this model initially to a limited group of critical industry partners and open source developers with Project Glasswing, we aim to enable defenders to begin securing the most important systems before models with similar capabilities become broadly available.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;In other words: &lt;strong&gt;Anthropic is trying to turn a capability advantage into a temporary defensive advantage before the rest of the ecosystem catches up&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;What is really changing: publishing no longer means deploying
 &lt;div id="what-is-really-changing-publishing-no-longer-means-deploying" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-is-really-changing-publishing-no-longer-means-deploying" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The most interesting thing about Mythos is not only the security argument. It is the precedent it sets.&lt;/p&gt;
&lt;p&gt;For years, many of us assumed that the most advanced model in a lab would also, sooner or later, be the one that ended up in the hands of customers, developers, or end users. With Mythos, that equivalence breaks.&lt;/p&gt;
&lt;p&gt;From now on, the most advanced model may:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;not be the main product,&lt;/li&gt;
&lt;li&gt;not be broadly offered via API,&lt;/li&gt;
&lt;li&gt;not reach the general market,&lt;/li&gt;
&lt;li&gt;and exist for some time in a kind of &lt;strong&gt;strategic quarantine&lt;/strong&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That changes a great deal.&lt;/p&gt;
&lt;p&gt;It changes how we think about competition between labs. It changes how we should read public announcements. And it changes the regulatory and geopolitical frame as well: &lt;strong&gt;if the most powerful models are no longer necessarily public, then the true frontier of capability may increasingly sit behind restricted-access programs, private agreements, and asymmetric deployments&lt;/strong&gt;.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;But a critical reading is still necessary
 &lt;div id="but-a-critical-reading-is-still-necessary" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#but-a-critical-reading-is-still-necessary" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;That said, it would be a mistake to swallow the narrative whole.&lt;/p&gt;
&lt;p&gt;Anthropic is making extraordinary claims:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;thousands of high-severity vulnerabilities,&lt;/li&gt;
&lt;li&gt;zero-days in critical software,&lt;/li&gt;
&lt;li&gt;coverage across every major OS and browser,&lt;/li&gt;
&lt;li&gt;sophisticated exploits developed autonomously,&lt;/li&gt;
&lt;li&gt;and a security rationale strong enough to justify withholding the model.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The problem is that &lt;strong&gt;the public evidence is necessarily limited&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Anthropic itself says that more than 99% of the vulnerabilities it has found are still unpatched and therefore cannot be disclosed. In addition, the risk document is presented in &lt;strong&gt;redacted&lt;/strong&gt; form.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“Over 99% of the vulnerabilities we’ve found have not yet been patched, so it would be irresponsible for us to disclose details about them.”&lt;br&gt;
— Anthropic, &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;That is reasonable from the standpoint of responsible disclosure. But it also means that much of this story depends on &lt;strong&gt;trusting the lab’s own interpretation and framing&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;So yes: Anthropic’s decision may be sensible, even admirable, while still being wrapped in a corporate narrative that deserves methodological skepticism.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;My read: Mythos may mark a before and after
 &lt;div id="my-read-mythos-may-mark-a-before-and-after" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#my-read-mythos-may-mark-a-before-and-after" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;My impression is that this episode may ultimately be remembered less for the model’s name than for the strategic signal it sends.&lt;/p&gt;
&lt;p&gt;Anthropic is not only saying “we trained something very powerful.” It is saying something more uncomfortable:&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;&lt;strong&gt;we have crossed a capability frontier where responsible behavior no longer automatically means publication&lt;/strong&gt;.&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;If that thesis holds, Mythos will matter for three reasons.&lt;/p&gt;

&lt;h3 class="relative group"&gt;1. Because it normalizes partial retention of frontier models
 &lt;div id="1-because-it-normalizes-partial-retention-of-frontier-models" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#1-because-it-normalizes-partial-retention-of-frontier-models" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h3&gt;
&lt;p&gt;Not as an anecdotal exception, but as a legitimate governance tool.&lt;/p&gt;

&lt;h3 class="relative group"&gt;2. Because it shifts the debate from “what can the model do?” to “who should be allowed to use it, and when?”
 &lt;div id="2-because-it-shifts-the-debate-from-what-can-the-model-do-to-who-should-be-allowed-to-use-it-and-when" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#2-because-it-shifts-the-debate-from-what-can-the-model-do-to-who-should-be-allowed-to-use-it-and-when" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h3&gt;
&lt;p&gt;That is a fundamental change.&lt;/p&gt;

&lt;h3 class="relative group"&gt;3. Because it suggests that the real frontier of capability may already sit several steps ahead of what we see in product
 &lt;div id="3-because-it-suggests-that-the-real-frontier-of-capability-may-already-sit-several-steps-ahead-of-what-we-see-in-product" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#3-because-it-suggests-that-the-real-frontier-of-capability-may-already-sit-several-steps-ahead-of-what-we-see-in-product" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h3&gt;
&lt;p&gt;And that has major implications for strategy, technology policy, and security.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;The uncomfortable conclusion
 &lt;div id="the-uncomfortable-conclusion" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-uncomfortable-conclusion" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;For years, the dominant AI narrative assumed that technical progress would eventually democratize access to ever more powerful capabilities.&lt;/p&gt;
&lt;p&gt;Claude Mythos introduces a different possibility: that some capabilities are so sensitive that technical progress will not lead to openness, but to &lt;strong&gt;containment&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Not because the model failed. Precisely because it worked too well.&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;“Claude Mythos Preview reveals a stark fact: AI models have reached a level of coding capability where they can surpass all but the most skilled humans at finding and exploiting software vulnerabilities.”&lt;br&gt;
— Anthropic, &lt;em&gt;Project Glasswing&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;&lt;p&gt;If Anthropic is right, this is not simply another model launch. It is the moment when a frontier lab explicitly decided that &lt;strong&gt;its most advanced system should not behave like a normal product&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;And in this industry, that is a much bigger story than any benchmark.&lt;/p&gt;
&lt;hr&gt;

&lt;h2 class="relative group"&gt;Main sources
 &lt;div id="main-sources" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#main-sources" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;Anthropic — &lt;em&gt;Project Glasswing&lt;/em&gt;&lt;br&gt;
&lt;a href="https://www.anthropic.com/glasswing" target="_blank" rel="noreferrer"&gt;https://www.anthropic.com/glasswing&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Anthropic Frontier Red Team — &lt;em&gt;Claude Mythos Preview&lt;/em&gt;&lt;br&gt;
&lt;a href="https://red.anthropic.com/2026/mythos-preview/" target="_blank" rel="noreferrer"&gt;https://red.anthropic.com/2026/mythos-preview/&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Anthropic — &lt;em&gt;Alignment Risk Update: Claude Mythos Preview (Redacted)&lt;/em&gt;&lt;br&gt;
&lt;a href="https://www.anthropic.com/claude-mythos-preview-risk-report" target="_blank" rel="noreferrer"&gt;https://www.anthropic.com/claude-mythos-preview-risk-report&lt;/a&gt;&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/claude-mythos-unreleased-frontier-model/featured.svg"/></item><item><title>Claude Mythos: The Model That Made Cybersecurity Stocks Crash — And What It Tells Us About Where AI Is Heading</title><link>https://carlesabarca.com/posts/claude-mythos-cybersecurity/</link><pubDate>Sun, 29 Mar 2026 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/claude-mythos-cybersecurity/</guid><description>Anthropic&amp;rsquo;s leaked next-generation model isn&amp;rsquo;t just more powerful — it can find and exploit software vulnerabilities faster than human defenders. The implications go far beyond cybersecurity.</description><content:encoded>&lt;p&gt;Three days ago, a misconfigured CMS at Anthropic left roughly 3,000 internal assets publicly accessible. Among them: a draft blog post announcing their next-generation AI model. The name varies between two leaked drafts — &amp;ldquo;Mythos&amp;rdquo; and &amp;ldquo;Capybara&amp;rdquo; — but what matters isn&amp;rsquo;t the name. What matters is what it can do.&lt;/p&gt;
&lt;p&gt;And what it can do should make anyone in technology leadership stop and think very carefully.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What Leaked
 &lt;div id="what-leaked" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-leaked" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;On March 26, security researchers Roy Paz (LayerX Security) and Alexandre Pauwels (University of Cambridge) discovered the exposed documents. Anthropic acknowledged the leak as &amp;ldquo;human error&amp;rdquo; and confirmed the model is real.&lt;/p&gt;
&lt;p&gt;Here&amp;rsquo;s what we know:&lt;/p&gt;
&lt;p&gt;Claude Mythos is not Claude Opus 4.7. It&amp;rsquo;s not an incremental update. It&amp;rsquo;s a &lt;strong&gt;new tier above Opus&lt;/strong&gt; — Anthropic&amp;rsquo;s own words: &amp;ldquo;a new name for a new tier of model: larger and more intelligent than our Opus models, which were, until now, our most powerful.&amp;rdquo; Reports suggest roughly 10 trillion parameters, a 5-10x jump from previous frontier models.&lt;/p&gt;
&lt;p&gt;Training is complete. Select customers are already testing it.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Why Cybersecurity Stocks Crashed
 &lt;div id="why-cybersecurity-stocks-crashed" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#why-cybersecurity-stocks-crashed" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The morning after the leak, the market&amp;rsquo;s reaction was swift and brutal:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;iShares Cybersecurity ETF: &lt;strong&gt;-4.5%&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;CrowdStrike, Palo Alto Networks, Zscaler, SentinelOne: &lt;strong&gt;-6% each&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;Tenable: &lt;strong&gt;-9%&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;Bitcoin dropped to $66,000&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Why? Because the leaked draft describes Mythos as &amp;ldquo;currently far ahead of any other AI model in cyber capabilities.&amp;rdquo; It can discover and exploit software vulnerabilities at speeds that — Anthropic&amp;rsquo;s own assessment — &amp;ldquo;far outpace human defenders.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;Read that again. The company that built it is telling you that human cybersecurity teams can&amp;rsquo;t keep up with it.&lt;/p&gt;
&lt;p&gt;This isn&amp;rsquo;t hypothetical. Anthropic already caught a Chinese state-sponsored group using Claude Code to infiltrate approximately 30 organizations — tech companies, financial institutions, government agencies — before detection. Mythos reportedly makes that look like child&amp;rsquo;s play.&lt;/p&gt;
&lt;p&gt;Stifel analyst Adam Borg put it plainly: &amp;ldquo;Mythos is an order of magnitude more powerful, and compute-intensive, than any other frontier model on the market.&amp;rdquo;&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Rollout Strategy Tells You Everything
 &lt;div id="the-rollout-strategy-tells-you-everything" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-rollout-strategy-tells-you-everything" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Anthropic&amp;rsquo;s deployment approach is perhaps the most revealing signal:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;First access:&lt;/strong&gt; Not developers. Not enterprises. &lt;strong&gt;Cybersecurity organizations&lt;/strong&gt; — &amp;ldquo;giving them a head start in improving the robustness of their codebases against the impending wave of AI-driven exploits.&amp;rdquo;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;No public launch date.&lt;/strong&gt; They&amp;rsquo;re explicitly delaying broad release.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Cost problem acknowledged.&lt;/strong&gt; Anthropic says it&amp;rsquo;s &amp;ldquo;very expensive to serve&amp;rdquo; and they need to make it &amp;ldquo;much more efficient before any general release.&amp;rdquo;&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;When a company builds the most powerful AI model in the world and its first instinct is to hand it to defenders before attackers can get it — that&amp;rsquo;s not a product launch. That&amp;rsquo;s a controlled disclosure.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What Mythos Means Beyond Cybersecurity
 &lt;div id="what-mythos-means-beyond-cybersecurity" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-mythos-means-beyond-cybersecurity" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Let me be direct about what I think this represents.&lt;/p&gt;
&lt;p&gt;Mythos posts &amp;ldquo;dramatically higher scores&amp;rdquo; than Opus 4.6 on coding and academic reasoning benchmarks. Opus 4.6 already led SWE-bench Verified at 80.8% and Terminal-Bench 2.0 at 65.4%. Whatever &amp;ldquo;dramatically higher&amp;rdquo; means, we&amp;rsquo;re talking about a model that can code better than most professional developers and reason through complex problems at a level that was science fiction five years ago.&lt;/p&gt;
&lt;p&gt;But the cybersecurity capability is the real wake-up call, because vulnerability discovery requires something qualitatively different from text generation or code completion. It requires:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Deep multi-step reasoning&lt;/strong&gt; — chaining logical inferences across complex systems&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Adversarial creativity&lt;/strong&gt; — finding attack vectors that weren&amp;rsquo;t designed or anticipated&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Autonomous execution&lt;/strong&gt; — not just identifying a vulnerability but actively exploiting it&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;When a model can do all three at superhuman speed in a domain as complex as cybersecurity, the implications extend to every field that involves complex reasoning under uncertainty. Law. Medicine. Scientific research. Strategic planning. Finance.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The AGI Question (Which Is the Wrong Question)
 &lt;div id="the-agi-question-which-is-the-wrong-question" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-agi-question-which-is-the-wrong-question" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Is Mythos AGI? No. It doesn&amp;rsquo;t learn new tasks from minimal examples the way humans can. It has no persistent memory, no self-improvement loop, no autonomous goal-setting.&lt;/p&gt;
&lt;p&gt;But here&amp;rsquo;s what I think matters more: &lt;strong&gt;we may be past the point where the AGI label matters practically.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;A model that can autonomously find and exploit zero-day vulnerabilities — something that previously required teams of elite human researchers — changes the game regardless of whether we call it &amp;ldquo;general&amp;rdquo; intelligence. Narrow superintelligence in high-stakes domains is more immediately consequential than theoretical AGI.&lt;/p&gt;
&lt;p&gt;The fact that Anthropic itself is alarmed enough to delay general release and prioritize defensive deployment tells you where we are on the capability curve.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Competitive Context Makes It Worse
 &lt;div id="the-competitive-context-makes-it-worse" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-competitive-context-makes-it-worse" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Mythos doesn&amp;rsquo;t exist in isolation:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;OpenAI&lt;/strong&gt; has finished pretraining a new model codenamed &amp;ldquo;Spud&amp;rdquo; — expected within weeks.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Google DeepMind&lt;/strong&gt; just launched Gemini 3.1 for real-time multimodal processing.&lt;/li&gt;
&lt;li&gt;Both Anthropic and OpenAI are timing major releases ahead of planned &lt;strong&gt;IPOs later in 2026.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is an arms race with IPO pressure. The incentives to push capability boundaries are enormous and increasing. The incentives for caution are&amp;hellip; well, we just saw how Anthropic&amp;rsquo;s caution played out. A CMS misconfiguration, and the whole world knows.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What This Means for Institutions
 &lt;div id="what-this-means-for-institutions" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-this-means-for-institutions" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;For universities, for governments, for any organization making decisions about AI strategy:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The planning horizon just compressed.&lt;/strong&gt; If you were thinking about AI governance frameworks as a 2027-2028 initiative, think again. Models with superhuman capabilities in specific domains are here now, not in a comfortable future.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Cybersecurity is no longer optional.&lt;/strong&gt; It&amp;rsquo;s existential. Every institution needs to assume that AI-powered attacks will become the norm, not the exception. The defenders need AI too — and they need it first.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The talent equation is shifting.&lt;/strong&gt; When a model can outperform human cybersecurity experts, the value isn&amp;rsquo;t in the technical execution — it&amp;rsquo;s in the judgment about when and how to deploy these capabilities. We need people who understand both the technology and its implications.&lt;/p&gt;
&lt;p&gt;I keep coming back to the same conclusion I wrote in my &lt;a href="../../posts/from-seo-to-aeo/" &gt;previous post on AEO&lt;/a&gt;: digital transformation in 2026 means preparing institutions for a world where AI systems are colleagues, not tools. Mythos just made that statement feel uncomfortably literal.&lt;/p&gt;
&lt;p&gt;Jensen Huang said AGI has arrived. He was wrong about the definition but right about the urgency. Whether we call it AGI or narrow superintelligence or just &amp;ldquo;really powerful AI&amp;rdquo; — the systems are here, they&amp;rsquo;re real, and the time to prepare was yesterday.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;Carles Abarca is Vice President of Digital Transformation at Tecnológico de Monterrey.&lt;/em&gt;&lt;/p&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/claude-mythos-cybersecurity/featured.jpg"/></item><item><title>The Chart That Predicts Which Jobs AI Will Kill (And They're Not the Ones You Think)</title><link>https://carlesabarca.com/posts/ai-jobs-displacement-anthropic/</link><pubDate>Mon, 09 Mar 2026 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/ai-jobs-displacement-anthropic/</guid><description>An Anthropic study analyzing 2 million conversations reveals the gap between what AI CAN do and what it IS doing. That gap is the coming tsunami.</description><content:encoded>&lt;p&gt;Look at this chart carefully. It&amp;rsquo;s not an analysis of what AI has destroyed. It&amp;rsquo;s an &lt;strong&gt;X-ray of what it&amp;rsquo;s about to destroy&lt;/strong&gt;.&lt;/p&gt;

&lt;figure&gt;
 &lt;img
 class="my-0 rounded-md"
 loading="lazy"
 decoding="async"
 fetchpriority="auto"
 alt=""
 width="1280"
 height="853"
 src="../../posts/ai-jobs-displacement-anthropic/featured_hu_7a629d8976dfc179.png"
 srcset="../../posts/ai-jobs-displacement-anthropic/featured_hu_7a629d8976dfc179.png 800w,/posts/ai-jobs-displacement-anthropic/featured_hu_d58629eebe18c99c.png 1280w"
 sizes="(min-width: 768px) 50vw, 65vw"
 data-zoom-src="../../posts/ai-jobs-displacement-anthropic/featured.png"
 /&gt;
 
 &lt;figcaption&gt;Source: Anthropic — Labor market impacts of AI (March 2026)&lt;/figcaption&gt;
 &lt;/figure&gt;
&lt;p&gt;The blue area is what AI &lt;strong&gt;can&lt;/strong&gt; do today. The red area is what AI &lt;strong&gt;is&lt;/strong&gt; doing today. The difference between them isn&amp;rsquo;t a safety margin. It&amp;rsquo;s a tsunami that hasn&amp;rsquo;t hit shore yet.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Study: 2 Million Conversations with Claude
 &lt;div id="the-study-2-million-conversations-with-claude" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-study-2-million-conversations-with-claude" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Anthropic just published &lt;a href="https://www.anthropic.com/research/labor-market-impacts" target="_blank" rel="noreferrer"&gt;Labor market impacts of AI: A new measure and early evidence&lt;/a&gt;, and it&amp;rsquo;s the most rigorous analysis I&amp;rsquo;ve seen on AI&amp;rsquo;s real employment impact.&lt;/p&gt;
&lt;p&gt;What did they do? They crossed three data sources:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;The &lt;strong&gt;O*NET database&lt;/strong&gt;, cataloging tasks across ~800 US occupations.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Real Claude usage data&lt;/strong&gt; — 2 million conversations analyzed via the Anthropic Economic Index.&lt;/li&gt;
&lt;li&gt;Theoretical estimates from &lt;strong&gt;Eloundou et al. (2023)&lt;/strong&gt; on which tasks an LLM can make at least twice as fast.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The result is a new metric: &lt;strong&gt;observed exposure&lt;/strong&gt; — not what AI could theoretically do, but what it&amp;rsquo;s actually doing in professional settings. And the most revealing finding isn&amp;rsquo;t the absolute numbers — it&amp;rsquo;s the &lt;strong&gt;gap&lt;/strong&gt; between the two.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The 10 Most Exposed Jobs
 &lt;div id="the-10-most-exposed-jobs" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-10-most-exposed-jobs" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The ranking won&amp;rsquo;t surprise anyone who&amp;rsquo;s been paying attention, but the numbers are brutal:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Computer Programmers — 75% coverage&lt;/strong&gt;. Three out of four tasks a programmer does, Claude already handles.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Customer Service Representatives&lt;/strong&gt;. First-party API traffic shows massive automation.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Data Entry Keyers — 67%&lt;/strong&gt;. Reading documents and entering data. The perfect automation use case.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The list continues: actuaries, financial analysts, technical writers. &lt;strong&gt;Office jobs. White-collar work. People with college degrees.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;On the other end, 30% of workers have &lt;strong&gt;zero exposure&lt;/strong&gt;. Cooks, motorcycle mechanics, lifeguards, bartenders. Jobs where hands, bodies, and physical context are irreplaceable.&lt;/p&gt;
&lt;p&gt;Ironic, isn&amp;rsquo;t it? Decades telling us automation was coming for manual labor. &lt;strong&gt;It&amp;rsquo;s coming for the desks.&lt;/strong&gt;&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Demographic Surprise
 &lt;div id="the-demographic-surprise" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-demographic-surprise" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;This is where the study shatters the dominant narrative.&lt;/p&gt;
&lt;p&gt;The workers most exposed to AI are:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;16 percentage points more likely to be female&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;11 points more likely to be white&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Nearly twice as likely to be Asian&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Earn 47% more&lt;/strong&gt; than unexposed workers&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;17.4% hold graduate degrees&lt;/strong&gt; (vs. 4.5% in the unexposed group)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This &lt;strong&gt;isn&amp;rsquo;t&lt;/strong&gt; the displaced factory worker narrative. These are lawyers, analysts, programmers, university professors. The professional class that thought it was untouchable.&lt;/p&gt;
&lt;p&gt;When I say this will reshape social structure, I&amp;rsquo;m not exaggerating.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Gap IS the Prediction
 &lt;div id="the-gap-is-the-prediction" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-gap-is-the-prediction" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Go back to the chart. Look at the categories:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Computer &amp;amp; Math&lt;/strong&gt;: 94% theoretical capability, 33% actual use&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Legal&lt;/strong&gt;: ~85% theoretical, less than 15% observed&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Education&lt;/strong&gt;: ~70% theoretical, less than 15% observed&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Office &amp;amp; Admin&lt;/strong&gt;: 90% theoretical, a fraction of actual use&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That distance between blue and red isn&amp;rsquo;t comfort. &lt;strong&gt;It&amp;rsquo;s latency.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;It&amp;rsquo;s the time companies take to adopt, regulators to adapt, workflows to reconfigure. But the technology is already there. The model already knows how. The ecosystem just needs to catch up.&lt;/p&gt;
&lt;p&gt;And every month, the red area grows. Anthropic says it explicitly: &lt;em&gt;&amp;ldquo;As capabilities advance, adoption spreads, and deployment deepens, the red area will grow to cover the blue.&amp;rdquo;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;This isn&amp;rsquo;t speculative prediction. It&amp;rsquo;s an &lt;strong&gt;empirical observation with trajectory&lt;/strong&gt;.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What Changes with AI Agents
 &lt;div id="what-changes-with-ai-agents" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-changes-with-ai-agents" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Here&amp;rsquo;s the factor the study &lt;strong&gt;doesn&amp;rsquo;t&lt;/strong&gt; fully measure — because it didn&amp;rsquo;t exist at this scale when they collected the data.&lt;/p&gt;
&lt;p&gt;The study analyzes LLM usage — conversations with Claude. Chat interactions. A human asks, the AI answers. It&amp;rsquo;s the &lt;strong&gt;augmentation&lt;/strong&gt; model: AI helps you, you execute.&lt;/p&gt;
&lt;p&gt;But &lt;strong&gt;AI agents&lt;/strong&gt; are something else entirely. They don&amp;rsquo;t answer — they &lt;strong&gt;act&lt;/strong&gt;. They execute task chains autonomously. They navigate systems, make intermediate decisions, complete entire workflows without human intervention.&lt;/p&gt;
&lt;p&gt;What we&amp;rsquo;re building at Tecnológico de Monterrey with &lt;strong&gt;AgenTECs&lt;/strong&gt; is exactly this. Not a chatbot that helps you draft an email. An agent that manages the entire process: reads context, drafts, sends, follows up, escalates if there&amp;rsquo;s no response.&lt;/p&gt;
&lt;p&gt;When agents arrive at enterprise scale — and they&amp;rsquo;re already arriving — &lt;strong&gt;the red area in the chart will expand explosively&lt;/strong&gt;. Because you no longer need a human interacting with AI task by task. The agent covers the entire role.&lt;/p&gt;
&lt;p&gt;Think about the Legal category: 85% theoretical capability, &amp;lt;15% current use. What happens when an agent can review contracts, identify risk clauses, generate executive summaries, and prepare response drafts — all without a lawyer touching the keyboard? The 85% becomes the new floor, not the ceiling.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What to Do (Which Is Not Panic)
 &lt;div id="what-to-do-which-is-not-panic" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-to-do-which-is-not-panic" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;I&amp;rsquo;ve been saying the same thing for years: this isn&amp;rsquo;t about fear. It&amp;rsquo;s about &lt;strong&gt;preparation&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;When I wrote &lt;a href="https://www.linkedin.com/pulse/el-fin-del-desarrollador-carles-abarca-kxg2c/" target="_blank" rel="noreferrer"&gt;&amp;ldquo;El fin del desarrollador&amp;rdquo;&lt;/a&gt; on LinkedIn, the reaction was predictable: &amp;ldquo;exaggerated,&amp;rdquo; &amp;ldquo;developers will always be needed,&amp;rdquo; &amp;ldquo;AI can&amp;rsquo;t do X.&amp;rdquo; The same arguments I heard about TECgpt when we launched it and people said professors would never use it. Today we have &lt;strong&gt;over 60,000 active users&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;The metaphor I use is the &lt;strong&gt;orchestra conductor&lt;/strong&gt;. The value is no longer in playing the violin — it&amp;rsquo;s in knowing what music to perform, who plays what, and when to change the score. Future professionals don&amp;rsquo;t execute tasks — they &lt;strong&gt;orchestrate systems&lt;/strong&gt; that execute them.&lt;/p&gt;
&lt;p&gt;Specifically:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Massive upskilling, now&lt;/strong&gt;. Not &amp;ldquo;intro to AI&amp;rdquo; courses — real training on production tools.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Redefine roles&lt;/strong&gt;, don&amp;rsquo;t eliminate them. A lawyer who masters AI agents is worth more, not less. But a lawyer who only knows manual contract review has an expiration date.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Measure exposure&lt;/strong&gt; in your organization. Use Anthropic&amp;rsquo;s framework. Identify which tasks in each role an LLM can already perform. Design the transition before it&amp;rsquo;s imposed on you.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Create new roles&lt;/strong&gt; that don&amp;rsquo;t exist yet: AI orchestrators, agent prompt engineers, autonomous systems supervisors.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 class="relative group"&gt;The Bottom Line
 &lt;div id="the-bottom-line" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-bottom-line" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The BLS projects that the most exposed occupations under this metric will grow &lt;strong&gt;less&lt;/strong&gt; through 2034. For every 10 points of observed coverage, the growth projection drops 0.6 percentage points. This isn&amp;rsquo;t casual correlation — labor market analysts are seeing the same thing.&lt;/p&gt;
&lt;p&gt;And yet, Anthropic also finds that &lt;strong&gt;there&amp;rsquo;s no systematic increase in unemployment&lt;/strong&gt; in the most exposed professions. Yet.&lt;/p&gt;
&lt;p&gt;That&amp;rsquo;s the window. We&amp;rsquo;re in the moment between seeing the lightning and hearing the thunder. &lt;strong&gt;The bolt already struck.&lt;/strong&gt; The question isn&amp;rsquo;t whether the sound will arrive, but whether you&amp;rsquo;ll be ready when it does.&lt;/p&gt;
&lt;p&gt;Those who read this chart as &amp;ldquo;AI hasn&amp;rsquo;t affected employment much yet&amp;rdquo; are confusing latency with safety. Those who read it as &amp;ldquo;a structural labor market shift is coming and we need to act now&amp;rdquo;&amp;hellip; they&amp;rsquo;re the ones who&amp;rsquo;ll still be conducting the orchestra.&lt;/p&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/ai-jobs-displacement-anthropic/featured.png"/></item><item><title>Anthropic vs. the Pentagon: When AI Ethics Collides with Military Power</title><link>https://carlesabarca.com/posts/anthropic-pentagon-rupture/</link><pubDate>Tue, 03 Mar 2026 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/anthropic-pentagon-rupture/</guid><description>The rupture between Anthropic and the Pentagon over military use of Claude reveals a fundamental fracture in the AI industry: how far does the responsibility of technology creators extend?</description><content:encoded>&lt;p&gt;Last week we witnessed something unprecedented in the history of artificial intelligence: an AI company standing up to the Pentagon and saying &lt;strong&gt;&amp;ldquo;no.&amp;rdquo;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Anthropic, creator of Claude — the only AI model currently authorized on the U.S. federal government&amp;rsquo;s classified systems — rejected the final terms of a &lt;strong&gt;$200 million contract&lt;/strong&gt; with the Department of Defense. The consequences were immediate and brutal.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Red Line
 &lt;div id="the-red-line" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-red-line" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;The conflict boiled down to two non-negotiable points for Anthropic:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Mass surveillance of American citizens.&lt;/strong&gt; The Pentagon wanted to use Claude to analyze bulk-collected data: search histories, GPS movements, credit card transactions, even the questions you ask your favorite chatbot. All cross-referenced to build profiles.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Autonomous weapons.&lt;/strong&gt; Systems that select and engage targets without a human making the final call. The 2026 military budget allocates &lt;strong&gt;$13.4 billion&lt;/strong&gt; to these weapons alone.&lt;/p&gt;
&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Anthropic didn&amp;rsquo;t argue that such weapons shouldn&amp;rsquo;t exist. In fact, they offered to work directly with the Pentagon to improve their reliability. But they determined that current AI models &lt;strong&gt;aren&amp;rsquo;t reliable enough&lt;/strong&gt; to make lethal decisions autonomously. The risk of indiscriminate fire, civilian casualties, or even harm to American troops was, in their analysis, too high.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The False Solution of &amp;ldquo;Cloud vs. Edge&amp;rdquo;
 &lt;div id="the-false-solution-of-cloud-vs-edge" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-false-solution-of-cloud-vs-edge" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;During negotiations, a proposal emerged: keep Anthropic&amp;rsquo;s AI in the cloud, out of the weapons themselves. The models would synthesize intelligence before an operation but wouldn&amp;rsquo;t make kill decisions. The AI&amp;rsquo;s hands would stay clean.&lt;/p&gt;
&lt;p&gt;Anthropic rejected this with a devastating technical argument: &lt;strong&gt;in modern military AI architectures, the distinction between cloud and edge no longer exists.&lt;/strong&gt; Drones operate through mesh networks connected to data centers. The Pentagon actively works to push computing closer to the battlefield. If a model in an AWS server in Virginia is making combat decisions, ethically there&amp;rsquo;s no difference from it being inside the drone.&lt;/p&gt;

&lt;h2 class="relative group"&gt;The Response: The Hammer
 &lt;div id="the-response-the-hammer" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#the-response-the-hammer" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;When Anthropic held its ground, the response was swift:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Trump ordered&lt;/strong&gt; all federal agencies to cease using Anthropic&amp;rsquo;s technology.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Pete Hegseth&lt;/strong&gt; (Defense Secretary) designated Anthropic a &lt;strong&gt;supply chain risk to national security&lt;/strong&gt;, barring any military contractor from doing business with the company.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;OpenAI announced&lt;/strong&gt; a Pentagon deal just hours later.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The message was clear: &lt;em&gt;play by our rules, or we destroy you.&lt;/em&gt;&lt;/p&gt;

&lt;h2 class="relative group"&gt;What Sam Altman Didn&amp;rsquo;t Explain
 &lt;div id="what-sam-altman-didnt-explain" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-sam-altman-didnt-explain" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Here&amp;rsquo;s the most troubling part. Days before the collapse, Sam Altman had publicly declared that OpenAI would also refuse to let its models be used in autonomous weapons. Solidarity with Anthropic.&lt;/p&gt;
&lt;p&gt;But while making those statements, &lt;strong&gt;he was already negotiating with the Pentagon.&lt;/strong&gt; And he closed the deal hours after Anthropic&amp;rsquo;s fall, ensuring his AI would only be deployed &amp;ldquo;in the cloud&amp;rdquo; — exactly the solution Anthropic dismissed as insufficient.&lt;/p&gt;
&lt;p&gt;Nearly 100 OpenAI employees signed an open letter supporting the same red lines as Anthropic. Altman will have to explain on Monday why what Anthropic rejected on principle, he accepted for business.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What&amp;rsquo;s Really at Stake
 &lt;div id="whats-really-at-stake" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#whats-really-at-stake" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;This crisis transcends a contract. It reveals three fundamental fractures:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;1. AI as a geopolitical weapon.&lt;/strong&gt; AI technology is no longer just a commercial product. It&amp;rsquo;s a strategic military asset, and governments are willing to use their full power to control it.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;2. The illusion of self-regulation.&lt;/strong&gt; Anthropic tried to set ethical limits from within. The response was a national security risk designation. What company will dare say &amp;ldquo;no&amp;rdquo; after this?&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3. The gap between words and action.&lt;/strong&gt; OpenAI talked principles and signed a check. It&amp;rsquo;s not the first time, and the industry should take note.&lt;/p&gt;

&lt;h2 class="relative group"&gt;My Take
 &lt;div id="my-take" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#my-take" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;I&amp;rsquo;ve spent over 20 years in technology, and I&amp;rsquo;ve seen many inflection points. This is one of them.&lt;/p&gt;
&lt;p&gt;Anthropic did something extraordinarily rare in the tech industry: &lt;strong&gt;sacrifice $200 million and their federal government access for an ethical position.&lt;/strong&gt; We can debate whether it was a smart business decision, but we can&amp;rsquo;t deny it was brave.&lt;/p&gt;
&lt;p&gt;What concerns me isn&amp;rsquo;t Anthropic — they&amp;rsquo;ll survive. What concerns me is the precedent. If an AI company that puts ethical limits on its technology can be designated a &amp;ldquo;national security risk,&amp;rdquo; we&amp;rsquo;re building a system where the only option is blind obedience.&lt;/p&gt;
&lt;p&gt;And obedient AI without restrictions, in the hands of unchecked power, is exactly the scenario that every AI safety researcher has been warning about for years.&lt;/p&gt;
&lt;p&gt;The question is no longer whether AI will transform warfare. &lt;strong&gt;The question is who decides the limits.&lt;/strong&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;What do you think of Anthropic&amp;rsquo;s stance? Principles or naivety? I&amp;rsquo;d love to hear your perspective on &lt;a href="https://linkedin.com/in/abarca/" target="_blank" rel="noreferrer"&gt;LinkedIn&lt;/a&gt; or &lt;a href="https://x.com/carlesabarca" target="_blank" rel="noreferrer"&gt;X&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/anthropic-pentagon-rupture/featured.png"/></item></channel></rss>