Talk on bias mitigation

TAS group members have organised an invited talk by Mateo Espinoza Zarlenga (University of Cambridge) on “Bias Mitigation in the Wild: Challenges and Opportunities”. The talk will be held on Wednesday 2 October 2024 in room D1.02 at 14:30.

The event is co-organised with the Computer System Institute at USI Faculty of Informatics and is also promoted by the recently-born Swiss XAI Cluster, which is a group of junior researchers working in explainable artificial intelligence across different Swiss universities.

Please find below the abstract of the talk and the short biography of the speaker.


Abstract:
Deep neural networks trained via empirical risk minimisation often exhibit significant performance disparities across groups, particularly when group and task labels are spuriously correlated (e.g., “grassy background” and “cows”). In this talk, I will first argue that previously proposed bias mitigation methods that aim to address this issue often either rely on group labels for training or validation, or require an extensive hyperparameter search, even when this assumption is not explicitly made. Such data and computational requirements hinder the practical deployment of these methods, especially when datasets are too large to be group-annotated, computational resources are limited, and models are trained through already complex pipelines. With this in mind, I will outline some of the challenges that may need to be addressed to design practical bias mitigation methods. Then, I will describe Targeted Augmentations for Bias mitigation (TAB), a new approach that takes these design principles into consideration. I will conclude by showing how TAB, a simple hyperparameter-free framework that leverages the entire training history of a helper model to identify spurious samples, improves worst-group performance without any group information or model selection.

Short Bio:
Mateo Espinoza Zarlenga is a final-year PhD student and Gates scholar at the University of Cambridge. His research is focused on developing interpretable deep learning methods that explain their predictions using human-understandable concept representations. Mateo has published papers at NeurIPS, ICML, AAAI, AIES, TMLR, ECCV, and ASPLOS. He holds an MPhil from the University of Cambridge and an MEng and BA from Cornell University. Webpage: https://hairyballtheorem.com/about/.


Posted

in

by

Tags:

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *