IAPS researchers Joe O'Brien Renan Araujo and Oliver Guest contributed to this memo in collaboration with the Oxford Martin AI Governance Initiative. The memo explores frameworks and criteria for determining which actors (e.g., government agencies, AI companies, third-party organizations) are best suited to develop AI model evaluations. Read more: https://bit.ly/3DYsUgL
New Research memo! Who Should Develop Which AI Evaluations? In the rapidly advancing field of AI, model evaluations are critical for ensuring trust, safety, and accountability. But who should be responsible for developing these evaluations? Our latest research explores the challenges include: 1. Conflicts of interest when AI companies assess their own models 2. The information and skill requirements for AI evaluations 3. The blurred boundary between developing and conducting evaluations To tackle these challenges, our researchers propose a taxonomy of four development approaches and present nine criteria for selecting evaluation developers, which we apply in a two-step sorting process to identify capable and suitable developers. Lara Thurnherr Robert Trager Christoph Winter Amin Oueslati Clíodhna Ní Ghuidhir Anka Reuel Merlin Stein Oliver Guest Oliver Sourbut Renan Araujo Yi Zeng Joe O'Brien Jun Shern Chan Lorenzo Pacchiardi Seth Donoghue Oxford Martin School Read the full report here: https://lnkd.in/etHrqCms