Research

Our Work

We do technical AI governance research, and provide guidance and comments to policymakers. We aim to answer technical questions to help us better govern advanced AI, this includes research into verification mechanisms for AI agreements, proposals to improve regulation, and emergency preparedness.

Research Categories

Technical Reports

Dec 2, 2024

What AI evaluations for preventing catastrophic risks can and cannot do

We analyze what AI evaluations can and cannot do for preventing catastrophic risks. While they can help determine lower bounds on AI capabilities, evaluations face fundamental limitations and cannot be solely relied upon to ensure safety.

Technical Reports

Nov 27, 2024

Mechanisms to Verify International Agreements About AI Development

In this research report we provide an in-depth overview of the mechanisms that could be used to verify adherence to international agreements about AI development.

Policy Comments

Oct 11, 2024

Response to BIS AI Reporting Requirements RFC

We respond to the BIS Request for Comment on the Proposed Rule for Establishment of Reporting Requirements for the Development of Advanced Artificial Intelligence Models and Computing Clusters.

Technical Reports

Sep 16, 2024

Declare and Justify: Explicit assumptions in AI evaluations are necessary for effective regulation

In this short paper we identify key assumptions in using evaluations to prevent harm from AI systems, and propose that regulation should require developers to adequately justify these assumptions.

Policy Comments

Sep 9, 2024

Comments on US AISI Managing Misuse Risk for Dual-Use Foundation Models (NIST AI 800-1)

We provide comments on the document Managing Misuse Risk for Dual-Use Foundation Models from US AISI. This guidance should emphasize the uncertainty in AI capability evaluations, and use less optimistic language.

Policy Comments

Jun 1, 2024

Comments on NIST AI RMF: Generative Artificial Intelligence Profile (NIST AI 600-1)

We provide comments on the NIST AI Risk Management Framework profile for generative AI. We suggest including risks from misaligned AI systems, and offer several concrete actions which could be added to the profile.

Policy Comments

May 17, 2024

Response to Romney, Reed, Moran, King Framework to Mitigate AI-Enabled Extreme Risks

We offer comments on the Framework for Mitigating Extreme AI Risks, from Senators Romney, Reed, Moran, and King. The framework is an excellent first step, and could be strengthened by including risks from autonomous AI systems.

Policy Comments

Apr 29, 2024

Response to OMB Procurement of Artificial Intelligence RFI

We respond to the OMB Request for Information on Responsible Procurement of Artificial Intelligence in Government. We urge the OMB to be forward-looking, so today’s policies can address the challenges of future AI systems.

Policy Comments

Mar 27, 2024

Response to NTIA AI Open Model Weights RFC

We respond to the NTIA Request for Comment on Dual Use Foundation Artificial Intelligence Models With Widely Available Model Weights. We highlight the need to set a good precedent for future AI systems.