Anthropic’s Security model
In the last couple of months, it’s been clear that agentic AI will automate broad classes of software development, and this week Anthropic announced a new model, Mythos, that demonstrates the other side of this - Mythos is very good at finding security exploits automatically. In particular, it’s good at searching across many different parts of a complex system to find multiple different weak spots that can be exploited and chained together to break in.
I often compare AI to interns (or associates), and that’s a comparison in two parts: the first part is to imagine you have a million interns, and so now you can have an intern listen to every customer call and tell you if the customer is angry; but the second kind is to imagine you have one intern who can listen to a million customer calls at once. That kind of ‘intern’ can give you insights that no human ever could, not because they’re more ‘clever’ but because of that scale. Mythos is that intern: a security researcher that can check everything all at once, and remember everything and make every connection.
So, Anthropic says that Mythos already found a whole bunch of exploits in lots of widely deployed software, including one in FreeBSD that’s been there for 27 years that would let anyone crash it. Accordingly, it’s not making Mythos publicly available yet, because it doesn’t want just anyone to get that kind of hacking capability - instead, it announced a partnership project (‘Project Glasswing’) with other major tech companies to use this to fix their systems. Anthropic has a track record of making apocalyptic and somewhat questionable statements, but this one is getting a lot more credence from security people, and it’s ironic that the US Department of Defence claims it’s a supply chain threat when it now claims to have the capability to hack anything on Earth, more or less. LINK, OPINION
Anthropic coding blows the doors off
Meanwhile, Anthropic did a deal to buy both TPU AI accelerator capacity from Google and buy its own chips for Broadcom. The announcement includes the fact that the company now has annualised revenue (past four weeks multiplied by 13) of $30bn, up from $9bn at the end of December 2025, $14bn in February, and $19bn in March. I don’t like this ‘annualising’, but on a monthly basis, that means the company has gone from ~$75m monthly revenue at the beginning of last year to close to $2.5bn last month.
Almost all of this is software development, where (see above and below), agentic coding uses vast amounts of tokens for vast amounts of money that software developers are willing to pay given the productivity gains. It is pretty widely reported that this inference itself is profitable, but there are lots of complaints that Anthropic is capacity-constrained (hence the infra deals) and meanwhile training the next model is a money pit (and each foundation model only remains relevant for six to nine months at best).
This revenue also puts Anthropic ahead of OpenAI, based on its last public claim of $2bn monthly revenue at the end of March ($24-25bn annualised). However, these are all rounded numbers, and more importantly, the two companies don’t recognise revenue in the same way: according to The Information a few weeks ago, when their models are resold by cloud partners, OpenAI records only its cut as revenue, whereas Anthropic records the total sale as revenue and then deducts the cloud providers’ share as a cost. LINK
For context, Amazon’s 2025 shareholder’s letter says AWS has $15bn in annualised AI revenue - i.e. ~$1.2bn monthly revenue. (The letter also points to Amazon’s launch of same-day delivery in broad parts of rural America, which would have been front-page news before ChatGPT.) LINK
Tokenmaxxing
The sheer amount of tokens you can consume if you’re spinning up squadrons of agents, automating dozens of tasks, can get intoxicating: someone at Meta started an internal leaderboard for which engineers were using the most - apparently over 20 days the total usage was 60tr, and the highest-ranked user had burnt 281 billion, which could cost anything from hundreds of thousands to millions of dollars. The board was taken down after it was reported. Meanwhile, Visa says its employees are using nearly 2tr a month. This reminds me a bit of showing off your mobile data bill (though at vastly greater scale), and some of it is just performative, but it’s also about how fast you are (or can say you are) adopting something that’s clearly transforming your industry, which is also a recruiting tool. LINK, VISA
Meta gets back into the game
It’s a measure of how much happened this week that I put this story in fourth place - Meta has released Muse Spark, the first model from the new AI lab that it built from scratch for billions of dollars last year. And, the model is pretty good - not right at the top of the league tables, but definitely in the pack with Anthropic, Google, and OpenAI (though we should be a little cautious, given that the Llama model from the previous team tried to game the benchmarks). The fact that Meta fell off the ladder and managed to get back on is a tribute to Mark Zuckerberg’s management skills, and also a rebuke to Apple, Amazon, and especially Microsoft, all of whom have failed to do the same. LINK
Sam Altman
The most terrifying sentence in the English language is “Rowan Farrow is writing a profile of you”, but this time he filled half the New Yorker with a profile of Sam Altman and didn’t find anything much that we don’t already know. Many people who know him say he’s an untrustworthy, manipulative liar, and many people who’ve worked with him quit, and whether you agree or not, all of that’s been written about at length already. Indeed, the only thing that was new to me was the note that someone with a grudge is spreading rumours that Altman pays for under-aged sex (which Farrow couldn’t find any support for). LINK
Meanwhile, someone threw a Molotov cocktail at Sam Altman’s house. The crazy fringes of the ‘doomers’ are still around, especially in the Bay Area, even as everyone else concluded they’re morons, but there’s also a growing panic about AI and data centres, sometimes rational but probably wrong (impact on employment), and sometimes based on pure misinformation (no, this isn’t ‘consuming’ lots of water). Either way, you would need to be a horrible and very stupid person to think any of that justifies violence. Sam Altman wrote about this here. LINK
PE/AI rollups
The WSJ reports that Anthropic is in talks with a range of PE funds to set up a new vehicle that would act as a consultant to help their portfolio companies deploy AI. Anthropic would invest $200m of a planned $1bn raise. There’s an obvious irony in the fact that deploying a thing that breathless Silicon Valley types think will destroy consultants turns out to need a lot of consulting. But as a company, why would you want a consultancy that’s captive to one vendor, especially when things are changing so fast? Meanwhile, Jeff Bezos’s new venture, that aims to use PE and AI to buy and transform companies, hired the last remaining co-founder out of Elon Musk’s xAI. LINK, BEZOS
In other news
The US tractor company John Deere, which has morphed into a tech company, settled a class-action ‘right to repair’ lawsuit. LINK
The New York Times did a big investigation claiming that an early cryptocurrency pioneer called Adam Back is the pseudonymous ‘Satyoshi Nakamoto’ who created Bitcoin. 🤷🏻♂️. LINK
Anthropic’s marketing team is hiring video editors at $250k each - one of many fields where AI is an accelerant, not a replacement. LINK
If you’re an ambulance-chasing lawyer looking for clients to sue Meta, Meta won’t let you use their platform for ads. There are many layers of irony here. LINK
|