UK financial regulators rush to assess risks of Anthropic’s latest AI model: report

1 minute, 18 seconds Read
British financial regulators are holding urgent talks with the government’s cybersecurity agency ​and major banks to assess risks posed by the ‌latest artificial intelligence model from Anthropic, the Financial Times reported on Sunday.

Officials at the Bank of England (BoE), the Financial Conduct Authority (FCA) and HM Treasury ​are in talks with the National Cyber Security Centre (NCSC) ​to examine potential vulnerabilities in critical IT systems highlighted ⁠by Anthropic’s latest AI model, the newspaper said.

Representatives from major ​British banks, insurers and exchanges are expected to be briefed ​on the cybersecurity risks posed by the model, Claude Mythos Preview, at a meeting with regulators in the next fortnight, it said, citing two ​people briefed on the talks.

Reuters could not immediately verify the ​report. Anthropic did not respond to Reuters’ request for comment, while the ‌BoE ⁠declined to comment, and the Treasury, NCSC and FCA were not immediately available for comment.

Read More: IBM to pay $17 million to settle US government probe over DEI

The move follows a meeting called by US Treasury Secretary Scott Bessent with major Wall Street banks on ​the model’s cyber ​risk potential, Reuters ⁠reported on Friday, citing two sources.

The AI startup has said the model is being deployed as ​part of “Project Glasswing”, a controlled initiative under which ​select ⁠organisations are permitted to use the unreleased Claude Mythos Preview model for defensive cybersecurity purposes.

In a blog post earlier this month, ⁠the ​startup said the model had already identified “thousands” ​of major vulnerabilities across operating systems, web browsers and other widely used software.

Similar Posts