Policy Options for Preserving Chain of Thought Monitorability
The most advanced AI models produce detailed reasoning steps in human language—known as "chain of thought" (CoT)—that provide crucial oversight capabilities for ensuring these systems behave as intended. However, competitive pressures may drive developers toward more efficient but non-monitorable architectures that lack a human-readable CoT. This report presents a framework for determining when coordination mechanisms are needed to preserve CoT monitorability.
Accelerating AI Data Center Security
AI systems are advancing at breakneck speed and already reshaping markets, geopolitics, and the priorities of governments. Frontier AI systems are developed and deployed using compute clusters of hundreds of thousands of cutting-edge AI chips housed in specialized data centers. These AI data centers are likely tempting targets for sophisticated adversaries like China and Russia, who may seek to steal intellectual property or sabotage AI systems underpinning military, industry, or critical infrastructure projects.
How AI Chips Are Made
Adapted from a section of a report by Erich Grunewald and Christopher Phenicie, this blog post introduces the core concepts and background information needed to understand the AI chip-making process.
Compute is a Strategic Resource
Computational power (“compute”) is a strategic resource in the way that oil and steel production capacity were in the past. Like oil, and like steel production capacity, compute is scarce, controllable, concentrated, and highly economically and militarily useful. Just as oil and steel were and remain strategic resources to some extent, compute is now also a strategic resource of very high importance.
The Hidden AI Frontier
The most advanced AI systems remain hidden inside corporate labs for months before public release—creating both America's greatest technological advantage and a serious security vulnerability. IAPS researchers identify critical risks and propose lightweight interventions to lessen the threat.
Promoting the Stack: Trump’s AI Export Incentive Program Explained
Alongside its AI Action Plan, the Trump administration published an executive order (EO) for Promoting the Export of the American AI Technology Stack.
Policy Actions for Enabling Cyber Defense Through Differential Access
In our Differential Access report, we provided a strategic framework to help developers give defenders an advantage by shaping access to AI-powered cyber capabilities. In a new policy memo, we outline government actions that can enable Differential Access and promote AI adoption for cyber defense.
Verification for International AI Governance
The growing impacts of artificial intelligence (AI) are spurring states to consider international agreements that could help manage this rapidly evolving technology. The political feasibility of such agreements can hinge on their verifiability—the extent to which the states involved can determine whether other states are complying. This report, published by the Oxford Martin School at the University of Oxford analyzes several potential international agreements and ways they could be verified.
IAPS Researchers React: The US AI Action Plan
The Trump Administration unveiled its comprehensive AI Action Plan on Wednesday. Experts at the Institute for AI Policy and Strategy reviewed the plan with an eye toward its national security implications. As AI continues to accelerate towards very powerful artificial general intelligence, our researchers discuss promising proposals for addressing critical AGI risks, offer key considerations for government implementation, and explore the plan's gaps and potential solutions.
Managing Risks from Internal AI Systems
The most powerful AI systems are used internally for months before they are released to the public. These internal AI systems may possess capabilities significantly ahead of the public frontier, particularly in high-stakes, dual-use areas like AI research, cybersecurity, and biotechnology. To address these escalating risks, this report recommends a combination of technical and policy solutions.
Response to the American Science Acceleration Project RFI
This post contains IAPS’s response to the Request for Information from Senators Heinrich and Rounds as part of the American Science Acceleration Project (ASAP), a national initiative to accelerate the pace of American technical innovation.
A National Center for Advanced AI Reliability and Security
This is a linkpost for a policy memo published by the Federation of American Scientists, which proposes scaling up a significantly enhanced “CAISI+” within the Department of Commerce.
How Some of China’s Top AI Thinkers Built Their Own AI Safety Institute
The emergence of the China AI Safety and Development Association (CnAISDA) is a pivotal moment for China’s frontier AI governance. How it navigates substantial domestic challenges and growing geopolitical tensions will shape conversations on frontier AI risks in China and abroad.
A Whistleblower Incentive Program to Enforce U.S. Export Controls
A Whistleblower Incentive Program to Enforce U.S. Export Controls: "A program modeled on the successful SEC program would help America overcome its export control enforcement woes.”
Countering AI Chip Smuggling Has Become a National Security Priority: An Updated Playbook for Preventing AI Chip Smuggling to the PRC
The Center for a New American Security (CNAS), in collaboration with the Institute for AI Policy and Strategy, has released a new working paper which catalogues evidence that substantial quantities of advanced artificial intelligence (AI) chips are being smuggled into China, undermining U.S. national security.
Accelerating R&D for Critical AI Assurance and Security Technologies
A memo to outline a strategic, coordinated policy approach supporting R&D to address urgent assurance and security challenges relating to frontier AI systems.
Asymmetry by Design: Boosting Cyber Defenders with Differential Access to AI
“Differential access” is a strategy to tilt the cybersecurity balance toward defense by shaping access to advanced AI-powered cyber capabilities. We introduce three possible approaches, Promote Access, Manage Access, and Deny by Default, with one constant across all approaches — even in the most restrictive scenarios, developers should aim to advantage cyber defenders.
Expert Survey: AI Reliability & Security Research Priorities
Our survey of 53 specialists across 105 AI reliability and security research areas identifies the most promising research prospects to guide strategic AI R&D investment.
Location Verification for AI Chips
Adding location verification features to AI chips could unlock new governance mechanisms for regulators, help enforce existing and future export controls by deterring and catching smuggling attempts, and enable post-sale verification of chip locations. This paper is meant to serve as an initial introduction to location verification use-cases for AI chips with comparison of different methods.