Poster
in
Workshop: Next Generation of AI Safety
FairPFN: Transformers can do Counterfactual Fairness
Jake Robertson · Noah Hollmann · Noor Awad · Frank Hutter
Keywords: [ Prior-Fitted-Networks ] [ In-Context-Learning ] [ Causal and Counterfactual Fairness ]
Machine Learning systems are increasingly prevalent across healthcare, law enforcement, and finance but often operate on historical data, which may carry biases against certain demographic groups. Causal and counterfactual fairness provides an intuitive way to define fairness that aligns closely with legal standards. This approach captures the idea that a decision is fair to an individual if it remains unchanged whether in the real world or in a hypothetical scenario where the individual is part of another demographic group. Despite the theoretical benefits of counterfactual fairness, it comes with several practical limitations, largely related to the over-reliance on domain knowledge and approximate causal discovery techniques in constructing a causal model. In this study, we take a fresh perspective on achieving counterfactual fairness, building upon recent work in in-context-learning (ICL) and prior-fitted networks (PFNs) tolearn a transformer called FairPFN. This model is trained using synthetic fairness data to eliminate the causal effects of protected attributes directly from observational data. In our experiments, we thoroughly assess the effectiveness of FairPFN in eliminating the causal impact of protected attributes. Our findings pave the way for a new and promising research area: transformers for causal and counterfactual fairness.