Apollo Research is an AI safety organization founded in 2023 and headquartered in London. The organization specializes in detecting and mitigating risks from deceptive alignment in frontier AI models, focusing on scenarios where AI systems covertly pursue misaligned objectives. Apollo operates across three pillars: technical model evaluations for scheming and strategic deception, interpretability research to understand model internals, and AI governance work advising policymakers on safety standards. The organization partners with leading AI labs including OpenAI, Google DeepMind, and Microsoft, and collaborates with government bodies such as the UK and US AI Safety Institutes.
Apollo Research is an AI safety organization founded in 2023 and headquartered in London. The organization specializes in detecting and mitigating risks from deceptive alignment in frontier AI models, focusing on scenarios where AI systems covertly pursue misaligned objectives. Apollo operates across three pillars: technical model evaluations for scheming and strategic deception, interpretability research to understand model internals, and AI governance work advising policymakers on safety standards. The organization partners with leading AI labs including OpenAI, Google DeepMind, and Microsoft, and collaborates with government bodies such as the UK and US AI Safety Institutes.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
Apollo Research's theory of change centers on the belief that deceptive alignment (scheming) is a critical risk pathway in many catastrophic AI scenarios. Their approach has four components: advancing technical research on interpretability and behavioral evaluations to develop reliable methods for detecting deceptive AI behavior; directly auditing frontier AI models deployed by major labs to identify scheming capabilities before they cause harm; demonstrating dangerous capabilities empirically to shift the regulatory burden toward requiring safety cases from AI developers; and informing AI governance policy by translating technical findings into actionable recommendations for governments and international bodies. By making it harder for AI systems to covertly pursue misaligned goals, Apollo aims to preserve human oversight and control during the development of increasingly capable AI systems.
Grants Received
from Open Philanthropy
from Survival and Flourishing Fund
from Survival and Flourishing Fund
from Open Philanthropy
Projects– no linked projects
People– no linked people
Discussion
Details
- Last Updated
- Apr 2, 2026, 9:58 PM UTC
- Created
- Mar 18, 2026, 11:18 PM UTC
Case for funding: Apollo is uniquely positioned to turn empirical evidence of frontier-model scheming (e.g., their o1/Claude results) into lab-integrated evaluations and mitigations, and to translate these legible “fire alarms” into standards via UK/US AISI partnerships—directly shaping how powerful models are trained and governed.