Efficient and Interpretable Transformer for Counterfactual Fairness

arXiv cs.LG / 4/30/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a key tension in high-stakes regulated domains: models must balance predictive accuracy with interpretability and counterfactual fairness requirements.
  • It proposes the Feature Correlation Transformer (FCorrTransformer), an attention-light transformer for tabular data whose attention matrix can be interpreted statistically as pairwise feature dependencies.
  • The method introduces Counterfactual Attention Regularization (CAR) to enforce group-invariant fair representations of sensitive features at the attention level, aiming for counterfactually fair predictions without explicit causal assumptions.
  • Experiments on imbalanced classification and regression benchmarks show that FCorrTransformer + CAR delivers strong counterfactual fairness, competitive predictive performance, and lower model complexity than standard transformer baselines.
  • Overall, the work presents a practical bridge between fairness theory and deployable model architectures for “responsible AI” in regulatory-sensitive industries.

Abstract

The growing reliance of machine learning models in high-stakes, highly regulated domains such as finance and insurance has created a growing tension between predictive performance, interpretability, and regulatory fairness requirements. In these settings, models are expected not only to deliver reliable predictions but also to provide transparent decision rationales and comply with strict fairness requirements. Attention-based transformers offer powerful mechanisms for modeling complex data relationships as demonstrated in various language tasks, yet their attention mechanisms alone do not ensure counterfactually fair predictions, even when combined with fairness-aware techniques. To address these limitations, we propose the Feature Correlation Transformer (FCorrTransformer), an attention-light architecture tailored for tabular data. In this design, the attention matrix admits a direct statistical interpretation as pairwise feature dependencies, enhancing both interpretability and efficiency. Leveraging this structure, we introduce Counterfactual Attention Regularization (CAR), a framework that enforces group-invariant fair representations of sensitive features at the attention level, promoting counterfactually fair predictions without relying on explicit causal assumptions. Empirical evaluations on imbalanced classification and regression benchmarks demonstrate that FCorrTransformer combined with CAR achieves strong counterfactual fairness while maintaining competitive predictive performance and substantially reducing model complexity compared with standard transformer-based baselines. Overall, this work bridges a critical gap between fairness theory and machine learning models, offering a practical framework for responsible AI in regulatory-sensitive domains.