Anthropic’s refusal to arm AI is exactly why the UK wants it


The Anthropic UK expansion story is less about diplomatic courtship and more about what happens when a government punishes a company for having principles. In late February, US Defence Secretary Pete Hegseth gave Anthropic CEO Dario Amodei a stark ultimatum: remove guardrails preventing Claude from being used for fully autonomous weapons and domestic mass surveillance, or face consequences. 

Amodei didn’t budge. He wrote that Anthropic could not “in good conscience” grant the Pentagon’s request, arguing that some uses of AI “can undermine rather than defend democratic values.” Washington’s response was swift. 

Trump directed every federal agency to immediately cease all use of Anthropic’s technology, and the Pentagon designated the company a supply chain risk, a label ordinarily reserved for adversarial foreign entities like Huawei. The US$200 million Pentagon contract was pulled. 

Defence tech companies instructed employees to stop using Claude and switch to alternatives. London, watching all of this unfold, saw something different.

The UK’s pitch

Staff at the UK’s Department for Science, Innovation and Technology (DSIT) have drawn up proposals for the US$380 billion company, ranging from a dual stock listing on the London Stock Exchange to an office expansion in the capital, according to multiple people with knowledge of the plans. Prime Minister Keir Starmer’s office has backed the effort, which will be put to Amodei when he visits in late May. 

Anthropic already has around 200 employees in Britain and appointed former prime minister Rishi Sunak as a senior adviser last year. The infrastructure for a meaningful UK presence is already there. What the British government is now offering is an explicit signal that Anthropic’s approach to AI–built on embedded ethical constraints–is an asset, not an obstacle.

A dual listing in London, if it materialised, would give Anthropic access to European institutional investors at a moment when its domestic regulatory standing remains under active legal challenge. The Pentagon’s appeal of the court-ordered injunction blocking the supply chain designation is still before the Ninth Circuit, and the outcome remains uncertain.

Ethics as a competitive advantage

The dispute has been framed largely as a legal and political fight. But its implications for global AI governance run deeper. Anthropic’s lawyers argued in court filings that Claude was not developed to be used for lethal autonomous weapons without human oversight, nor deployed to spy on US citizens, and that using the tools in these ways would represent an abuse of its technology. 

US District Judge Rita Lin, who granted a preliminary injunction blocking the blacklist in March, found the government’s actions “troubling” and concluded they likely violated the law. That judicial finding matters in the UK context. Britain is positioning itself as a regulatory environment sitting between Washington’s current posture, which demands unrestricted military access, and Brussels, where the EU AI Act imposes its own constraints. 

The UK government presents itself as offering a less constrained environment for AI companies than either the US or the European Union. Crucially, that pitch doesn’t ask Anthropic to abandon the guardrails it went to court to defend.

The courtship also sits alongside broader UK efforts to build domestic AI capability, including a recently announced £40 million state-backed research lab, after officials acknowledged the absence of a homegrown competitor to the leading US frontier labs.

Competition in London

The UK’s play for Anthropic is not happening in a vacuum. OpenAI has already committed to making London its biggest research hub outside the US. Google has anchored itself in King’s Cross since acquiring DeepMind in 2014. The race to secure frontier AI in London is already competitive, and Anthropic’s current circumstances make it the most consequential target yet.

Anthropic has been expanding internationally regardless of its domestic legal battles, including opening a Sydney office as its fourth Asia-Pacific location. The global growth strategy is already in motion. What remains to be seen is how much of it London gets to claim.

The company Washington blacklisted for having an AI ethics policy is now being actively courted by another G7 government that wants exactly that. The late May meetings with Amodei will be telling.

See Also: Anthropic selected to build government AI assistant pilot

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.