About the Talk:
Yonadav presents his research Practices for Governing Agentic AI Systems. Read the entire paper here.
Agentic AI systems—AI systems that can pursue complex goals with limited direct supervision—are likely to be broadly useful if we can integrate them responsibly into our society. While such systems have substantial potential to help people more efficiently and effectively achieve their own goals, they also create risks of harm. In this white paper, we suggest a definition of agentic AI systems and the parties in the agentic AI system life-cycle, and highlight the importance of agreeing on a set of baseline responsibilities and safety best practices for each of these parties. As our primary contribution, we offer an initial set of practices for keeping agents’ operations safe and accountable, which we hope can serve as building blocks in the development of agreed baseline best practices. We enumerate the questions and uncertainties around operationalizing each of these practices that must be addressed before such practices can be codified. We then highlight categories of indirect impacts from the wide-scale adoption of agentic AI systems, which are likely to necessitate additional governance frameworks.
About the Speaker:
Yonadav Shavit is a former AI researcher who now focuses full time on AI policy, and specifically on frontier AI safety and security issues. He has a PhD in Computer Science from Harvard University, where he studied mechanisms for enforceable international treaties on large-scale AI training, and previously led AI policy at Schmidt Futures.