Please enable javascript for this website.

Recent work

Pre-Deployment Evaluation of Anthropic’s Upgraded Claude 3.5 Sonnet

Research

November 19, 2024

The UK Artificial Intelligence Safety Institute and U.S. Artificial Intelligence Safety Institute conducted a joint pre-deployment evaluation of Anthropic’s latest model

Safety case template for ‘inability’ arguments

Research

November 14, 2024

How to write part of a safety case showing a system does not have offensive cyber capabilities

Announcing Inspect Evals

Research

November 13, 2024

We’re open-sourcing dozens of LLM evaluations to advance safety research in the field

Our mission is to equip governments with an empirical understanding of the safety of advanced AI systems.

Empirical research

Monitoring the fast-moving landscape of AI development

Evaluating the risks AI poses to national security and public welfare

Advancing the field of systemic safety to improve societal resilience

Global impact

Working with AI developers to ensure responsible development

Informing policymakers about current and emerging risks from AI

Promoting global coordination on AI governance

Join us to shape the trajectory of AI

For our ambitious and urgent mission, we need top talent. We have built a unique structure within the government so we can operate like a startup. We have recruited over 30 technical staff, including senior alumni from OpenAI, Google DeepMind and the University of Oxford, and we are scaling rapidly. Our staff are supported by substantial funding and computing resources, priority access to top models, partnerships with leading research organisations and an incredibly talented, close-knit and driven team.