Evaluating Prototype Explanations in Machine Learning

Prototype-based post-hoc explanations aim to make model predictions interpretable by presenting representative examples (prototypes) that illustrate how the model arrives at decisions. Their evaluation often relies on quantitative metrics such as fidelity (how closely prototypes approximate the model’s decision function), coverage (how much of the input space they represent), stability (whether explanations remain consistent under small perturbations), and diversity (ensuring prototypes capture … Read more

[Taken] Intersectional Fairness in Machine Learning

This project focuses on the rich field of algorithmic fairness where the goal is to ensure that predictions are not biased against subgroups of the population whilst maximising predictive performance. One challenge is when we focus on multiple protected attributes.