|Talks|

On the Statistical Foundations of Auditing and Accountability in Modern ML Systems

Visiting speaker
Hybrid
Past Talk
Himabindu Lakkaraju
Assistant Professor, Harvard University
Feb 27, 2026
2:00 pm
EST
Feb 27, 2026
2:00 pm
In-person
Portsoken Street
London, E1 8PH, UK
The Roux Institute
Room
100 Fore Street
Portland, ME 04101
Network Science Institute
2nd floor
Network Science Institute
11th floor
177 Huntington Ave
Boston, MA 02115
Network Science Institute
2nd floor
Room
58 St Katharine's Way
London E1W 1LP, UK
Register for talk
Register for talk

Talk recording

Modern machine learning systems are increasingly developed through multi-stage training pipelines—pretraining, fine-tuning, and alignment—and deployed in settings where failures can be costly. Despite their impressive performance, these systems can exhibit brittle or unsafe behaviors, raising a central accountability question: when a model fails, which components of the learning pipeline or input features are responsible, and how can we reliably diagnose and address them? In this talk, I will present our recent work that develops statistical foundations for auditing and accountability in modern ML systems at two complementary levels. First, auditing training pipelines (stage-level accountability): we formalize the effect of removing or altering a training stage as a causal estimand using the potential outcomes framework. We define counterfactual stage effects and derive retraining-free, efficient estimators that approximate these effects while accounting for realistic optimization dynamics, including learning-rate schedules, momentum, and weight decay. The resulting stage-effect representations enable scalable audits of model behavior across diverse tasks and evaluation metrics. We further provide explicit worst-case bounds on approximation error, yielding formal guarantees on the fidelity of these estimates. Second, auditing model behavior locally (feature-level accountability): I will show that many widely used feature attribution methods can be interpreted as instances of local function approximation, with key differences driven by their neighborhood distributions and loss functions. Building on this perspective, I will describe Bayesian formulations that quantify uncertainty in feature attributions and derive principled sampling strategies for reliable estimation. Finally, I will discuss prescriptive feature-level interventions (counterfactual explanations) and how to ensure their robustness when the underlying model evolves. I will conclude by outlining future directions for strengthening the theoretical and practical foundations of auditing and improving modern machine learning systems.
About the speaker
Himabindu (Hima) Lakkaraju is an Assistant Professor at Harvard University with appointments in the Business School and the Department of Computer Science. She is also a Senior Staff Research Scientist at Google. Her research focuses on the algorithmic and applied aspects of trustworthy AI and its implications for real-world decision-making. Before joining Harvard, Hima earned her PhD and master’s degrees in Computer Science from Stanford University and held positions at Microsoft Research, IBM Research, and Adobe. Her research has been recognized with numerous accolades, including an Alfred P. Sloan Research Fellowship in Computer Science, a National Science Foundation (NSF) CAREER Award, a Kavli Fellowship from the U.S. National Academy of Sciences, an AI2050 Early Career Fellowship from Schmidt Sciences, multiple best paper awards at leading conferences, and grants from Google, OpenAI, Microsoft, Meta, Amazon, JPMorgan, and Bayer. Hima has been named one of the world’s top innovators under 35 by both MIT Technology Review and Vanity Fair. She has delivered keynote talks at prominent conferences and workshops, including INFORMS Annual Meetings, ICML, NeurIPS, CIKM, Stanford HAI, and MIT EmTech, as well as at events organized by the National Institute of Standards and Technology (NIST), the U.S. Securities and Exchange Commission (SEC), the U.S. Department of Commerce, and the United Nations. Her work has also been featured in major media outlets such as The New York Times, TIME, and Forbes. More recently, Hima co-founded the Trustworthy ML and Regulatable ML Initiatives to provide accessible resources and foster a community around these critical topics.
Share this page:
Feb 27, 2026