First Progress Report Towards Ambitions of the AI Safety Institute
In the first progress report foundational AI safety research advances include the establishment of a specialised research team and expert advisory board, alongside partnerships with leading organisations (ARC Evals, Advai, The Centre for AI Safety, Collective Intelligence Project, Faculty, Gryphon Scientific, RAND, Redwood Research, and Trail of Bits).
The AISI has set three priority areas to achieve its ambitions, including evaluations of advanced AI models, conducting foundational AI Safety research and facilitating information exchange. These are the key commitments from the first progress report.
1) Foundational AI Safety research
-
Presented work plan and mission to build an AI research team that can evaluate risk at the frontier of AI with technical evaluations by a neutral third party
-
Presentation of the name change from Taskforce to Frontier AI taskforce
-
Established an expert advisory board spanning AI research and national security: Yoshua Bengio, Paul Christiano, Matt Collins, Anne Keast-Butler, Alex van Someren, Helen Stokes-Lampard and Matt Clifford
-
Growing the Taskforce team to include leading experts such as Yarin Cal and David Krueger.
2) Facilitating Information exchange
-
Initial partnerships with leading organisations with ARC Evals, Advai, The Centre for AI Safety, Collective Intelligence Project, Faculty, Gryphon Scientific, RAND, Redwood Research and Trail of Bits.
You can read more about the second and third progress reports and the ambitions of the institute.
If you would like to learn more, please email [email protected].