Skip to content

Search the site

Anthropic’s new AI model can control your computer - badly

"Since Claude can interpret screenshots from computers connected to the internet, it’s possible that it may be exposed to content that includes prompt injection attacks..."

Anthropic’s latest AI model, out now, can control your computer – “looking at a screen, moving a cursor, clicking, and typing text.”

The capability is available in the company’s upgraded Claude 3.5 Sonnet and new model, Claude 3.5 Haiku. It represents what Anthropic said was the “first frontier AI model to offer computer use in public beta.”

“We've built an API that allows Claude to perceive and interact with computer interfaces. This API enables Claude to translate prompts into computer commands. Developers can use it to automate repetitive tasks, conduct testing and QA, and perform open-ended research” the firm said.

See also: No LLMs aren’t about to “autonomously” hack your company


Initial behaviour is buggy, Anthropic cautioned today: “Claude 3.5 Sonnet's current ability to use computers is imperfect. Some actions that people perform effortlessly—scrolling, dragging, zooming—currently present challenges. So we encourage exploration with low-risk tasks.”

The company added in a pair of write-ups today: “On OSWorld, which evaluates AI models' ability to use computers like people do, Claude 3.5 Sonnet scored 14.9% in the screenshot-only category… nowhere near human-level skill (which is generally 70-75%), but it’s far higher than the 7.7% obtained by the next-best AI model in the same category.

“Since Claude can interpret screenshots from computers connected to the internet, it’s possible that it may be exposed to content that includes prompt injection attacks” it warned, sharing a reference implementation

Claude 3.5 Haiku meanwhile is the next generation of its fastest model and outperforms GPT-4o and the original Claude 3.5 Sonnet. 

It will be released later this month said Anthropic and be available via first-party API, Amazon Bedrock, and Google Cloud’s Vertex AI – initially, it added,  as a text-only model and with image input to follow. 

Pre-deployment testing was conducted by the US AI Safety Institute (US AISI) and the UK Safety Institute (UK AISI) the company added. 

More detail and reaction to follow. What are your views on the potential uses here? Early testers, Red Teamers,tinkerers, share thoughts

In other AI news, Radiohead singer Thom Yorke and author Kazuo Ishiguro have joined over 10,000 signatories to a campaign urging a ban on the “unlicensed use of creative works for training generative AI.” 

British composer Ed Newton-Rex started the campaign. 

He told The Guardian: “There are three key resources that generative AI companies need to build AI models: people, compute, and data. 

“They spend vast sums on the first two – sometimes a million dollars per engineer, and up to a billion dollars per model. 

“But they expect to take the third – training data – for free."

Anthropic is among those named in the many class action suits against AI companies on this issue; law firm Mishcon De Raya has a tracker here.

Latest