×
AI safety research gets $40M offering from Open Philanthropy
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Open Philanthropy has announced a $40 million grant initiative for technical AI safety research, with potential for additional funding based on application quality.

Program scope and structure: The initiative spans 21 research areas across five main categories, focusing on critical aspects of AI safety and alignment.

  • The research areas include adversarial machine learning, model transparency, theoretical studies, and alternative approaches to mitigating AI risks
  • Applications are being accepted through April 15, 2025, beginning with a 300-word expression of interest
  • The program is structured to accommodate various funding needs, from basic research expenses to establishing new research organizations

Key research priorities: The initiative emphasizes understanding and addressing potential risks in AI systems while improving their reliability and transparency.

  • Adversarial machine learning research will focus on jailbreaks, control evaluations, and alignment stress tests
  • Model transparency investigations will explore white-box techniques, activation monitoring, and feature representations
  • Studies will examine sophisticated misbehavior in Large Language Models (LLMs), including alignment faking and encoded reasoning
  • Projects exploring theoretical aspects will investigate inductive biases and approaches to aligning superintelligence

Grant flexibility and support: Open Philanthropy has designed the program to be inclusive and accessible to various research entities and funding needs.

  • Grant types include support for research expenses, discrete projects lasting 6-24 months, and academic start-up packages
  • Funding is available for both existing nonprofits and the establishment of new research organizations
  • The program encourages applications even from those uncertain about their project’s exact fit, maintaining a low barrier to entry

Application process: The initiative emphasizes accessibility and transparency in its application procedures.

  • Initial submissions require only a brief 300-word expression of interest
  • Detailed information about research areas, eligibility criteria, and example projects is available in the full Request for Proposals
  • Questions can be directed to aisafety@openphilanthropy.org

Future implications: This substantial funding initiative signals a growing recognition of the importance of AI safety research while potentially reshaping the landscape of technical AI safety development.

  • The program’s broad scope and significant funding could accelerate progress in understanding and addressing AI risks
  • The initiative’s experimental nature in gauging funding demand may influence future investment patterns in AI safety research
  • The diverse range of supported research areas suggests a comprehensive approach to addressing AI safety challenges
Open Philanthropy Technical AI Safety RFP - $40M Available Across 21 Research Areas

Recent News

AI courses from Google, Microsoft and more boost skills and résumés for free

As AI becomes critical to business decision-making, professionals can enhance their marketability with free courses teaching essential concepts and applications without requiring technical backgrounds.

Veo 3 brings audio to AI video and tackles the Will Smith Test

Google's latest AI video generation model introduces synchronized audio capabilities, though still struggles with realistic eating sounds when depicting the celebrity in its now-standard benchmark test.

How subtle biases derail LLM evaluations

Study finds language models exhibit pervasive positional preferences and prompt sensitivity when making judgments, raising concerns for their reliability in high-stakes decision-making contexts.