Been Kim

Senior Staff Research Scientist at Google Deepmind

Speaker

Towards Interpretability for Humanity.

Peter Henderson

Assistant Professor at Princeton University

Speaker

Aligning Machine Learning and Law for Responsible Real-World Deployments

Aligning Machine Learning and Law for Responsible Real-World Deployments Abstract: Machine learning (ML) is being deployed to a vast array of real-world applications with profound impacts on society. ML can have positive impacts, such as aiding in the discovery of new cures for diseases and improving government transparency and efficiency. But it can also be harmful: reinforcing authoritarian regimes, scaling the spread of disinformation, and exacerbating societal biases. As we rapidly move toward systemic use of ML in the real world, there are many unanswered questions about how to successfully use ML for social good while preventing its potential harms. Many of these questions inevitably require pursuing a deeper alignment between ML and law. Are certain algorithms truly compliant with current laws and regulations? Is there a better design that can make them more in tune to the regulatory and policy requirements of the real world? Are laws, policies, and regulations sufficiently informed by the technical details of ML algorithms, or will they be ineffective and out-of-sync? In this talk, I will discuss ways to bring together ML and law together to address these questions. I will draw on real-world examples throughout the talk, including unique real-world collaborations with public sector organizations. I will show how investigating questions of alignment between ML and law can advance core research in ML, as well as creating positive public good impact. It is my hope that the tools discussed in this talk will help us lead to more effective and responsible ways of deploying ML in the real world, so that we steer toward positive impacts of ML.

Zico Kolter

Associate Professor, Carnegie Mellon University

Speaker

LLM Robustness: Recent progress and the challenges ahead.
Rida Qadri

Rida Qadri

Research Scientist at Google

Speaker

AI's Cultural Futures: Designing for a Culturally Rich World

With global diffusion of AI models, concerns about cultural erasure and Western bias are growing. In this talk, I explore how we can reimagine AI design to celebrate, not erase, the world's cultural richness. My talk focuses on three key aspects of AI design: imagination, evaluation, and interaction. First, I advocate for community-based studies to understand the diverse ways people imagine and engage with AI. By recognizing how communities appropriate and innovate with technology, we can develop more culturally appropriate uses and designs. Second, I challenge the dominance of evaluation metrics that neglect the nuances of social behavior they are seeking to evaluate. Finally, I envision AI interfaces that encourage users to question, critique, and playfully subvert the technology. This fosters a more empowered relationship, allowing users to shape AI in ways that reflect their own values and needs. This talk contributes to a vision of technology design that empowers communities to shape their own technological futures.

Hannah Rose Kirk

Research Scientist at the UK AI Safety Institute

Speaker

A Tale of Two RCTs: ​ Building a rigorous evidence base on the societal impacts of frontier AI inside the UK Government

This talk explores how the UK AI Safety Institute's Societal Impacts team is developing empirical evidence on how frontier AI affects human behaviour and beliefs. Drawing from our sociotechnical safety framework—which examines exposure, vulnerability, and severity of AI risks—I'll present two parallel research tracks that combine population-level insights with randomised control trials. The first examines AI's role in political information seeking and belief change, while the second investigates the relationship between AI anthropomorphism, personalisation and behavioural influence. Together, these studies demonstrate build evidence on the mechanisms through which AI systems can influence their users.

Margaret Mitchell

Chief Ethics Scientist, Hugging Face

Panelist

Yoshua Bengio

Scientific Director of Mila - Quebec AI Institute

Panelist

Jeff Clune

Professor at University of British Columbia

Panelist

Jakob Foerster

Associate Professor, University of Oxford

Panel Moderator