AI Security Institute

British research organisation From Wikipedia, the free encyclopedia

The AI Security Institute (AISI) is a research organisation under the Department for Science, Innovation and Technology, UK, that aims "to equip governments with a scientific understanding of the risks posed by advanced AI".[1] It conducts research and develop and test mitigations. Previously, it was known as the AI Safety Institute.[2] Its creation followed world's first major AI Safety Summit that was held in Bletchley Park in 2023.[3] The institute's professed goal is "building the world's leading understanding of advanced AI risks and solutions, to inform governments so they can keep the public safe". It is designed like a startup in the government "combining the authority of government with the expertise and agility of the private sector".[4]

FormationNovember 2023; 2 years ago (2023-11)
ServicesAI safety research
Quick facts Formation, Services ...
AI Security Institute
FormationNovember 2023; 2 years ago (2023-11)
ServicesAI safety research
Parent organization
Department of Science, Innovation and Technology
Websitehttps://www.aisi.gov.uk/
Close

AISI has made access agreements with Anthropic, Google and OpenAI to test their models before release.[3] It has an open source platform called Inspect that permits companies, governments and academics to run standardised safety tests for AI usage.[3] Among the works AISI has done is the reported detection of multiple serious vulnerabilities that could enable development of biological weapons; the vulnerabilities were fixed before the model was launched.[3]

It conducts research on diverse fields of AI application. One study by AISI found that LLMs post-trained for political persuasiveness became systematically less accurate and up to 51% more persuasive on political issues.[5] AISI has also worked on the usage of AI for emotional needs. It found that nearly 10 percent of UK citizens used systems like chatbots for emotional purposes on a weekly basis.[6] It found that "systems are now outperforming PhD-level researchers on scientific knowledge tests and helping non-experts succeed at lab work that would previously have been out of reach" in a report published in December 2025.[7]

Former chief AI officer of GCHQ Adam Beaumont is the institution's interim director. UK prime minister's AI advisor Jade Leung is the chief technology officer.[4]

See also

References

Related Articles

Wikiwand AI