AI Lab Watch is a project created by Zach Stein-Perlman that collects safety recommendations for frontier AI companies and evaluates them on their implementation. The project maintains a weighted scorecard assessing companies like Anthropic, OpenAI, Google DeepMind, Meta, xAI, Microsoft, and DeepSeek across categories including risk assessment, scheming prevention, safety research, misuse prevention, security, risk information sharing, and planning. It also publishes blog posts analyzing company safety practices and maintains resource pages documenting company commitments and integrity incidents. The project is motivated by concern about future powerful AIs causing existential catastrophe.
AI Lab Watch is a project created by Zach Stein-Perlman that collects safety recommendations for frontier AI companies and evaluates them on their implementation. The project maintains a weighted scorecard assessing companies like Anthropic, OpenAI, Google DeepMind, Meta, xAI, Microsoft, and DeepSeek across categories including risk assessment, scheming prevention, safety research, misuse prevention, security, risk information sharing, and planning. It also publishes blog posts analyzing company safety practices and maintains resource pages documenting company commitments and integrity incidents. The project is motivated by concern about future powerful AIs causing existential catastrophe.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
AI Lab Watch operates on the theory that public accountability and transparency can pressure frontier AI companies to adopt better safety practices. By collecting what AI labs should do to prevent extreme risks (such as AI takeover and human extinction), publicly scoring them on their actual practices, and documenting their commitments and integrity, the project creates reputational incentives for companies to improve. The scorecard highlights specific actions companies can take in areas like dangerous capability evaluations, scheming prevention, and safety research, making it easier for external stakeholders to assess and compare company behavior. By also analyzing companies' model eval reports and safety claims, the project helps ensure companies cannot make misleading claims about their safety practices without scrutiny.
Grants Received
from Survival and Flourishing Fund
Projects– no linked projects
People– no linked people
Discussion
Sign in to join the discussion.
Key risk: The project is winding down with Zach not maintaining the site and not currently accepting funding, making impact fragile to single-operator continuity and raising counterfactual concerns that similar tracking could be reproduced by larger governance groups.
Details
- Last Updated
- Apr 2, 2026, 10:11 PM UTC
- Created
- Mar 18, 2026, 11:18 PM UTC
Case for funding: Fund AI Lab Watch because its rigorously weighted seven-category scorecard and companion claims analysis provide the most credible, independent synthesis of frontier labs’ safety practices, create public comparators that pressure labs to improve, and have already influenced external indices like FLI’s AI Safety Index.