Fair Decision Making

Feb 17, 2026 · 3 min read

Overview

The Fair Decision Making project addresses an important challenge in modern machine learning: generating high-quality synthetic data while ensuring fairness across different demographic groups. As machine learning models increasingly influence high-stakes decisions in healthcare, finance, and criminal justice, ensuring these systems make bias-free predictions has become paramount.

This research program develops novel deep generative model architectures and training methodologies that produce synthetic data exhibiting both high utility (quality, diversity, fidelity) and fairness properties as well as representation learning techniques that aids for fair decision making process. Our work primarily focuses on healthcare applications, where biased models can have serious real-world consequences for patient care and treatment decisions.

Motivation

Training machine learning models on real-world data often leads to biased outcomes due to:

  1. Spurious correlations between sensitive attributes (e.g., race, gender) and target variables
  2. Underrepresentation of minority subgroups in training data
  3. Historical biases embedded in data collection processes

Synthetic data generation offers a promising solution by allowing us to create de-identified, representative datasets that maintain utility while mitigating these fairness concerns. However, existing synthetic data generators often fail to adequately address fairness, leading to bias amplification in downstream tasks.

Research Approach

Our research tackles fairness in synthetic data generation through multiple approaches:

1. Bias-Transforming Generative Models

We develop generative adversarial networks (GANs) with specific fairness constraints that learn to transform biased data distributions into fair ones. This includes:

  • Information-constrained data generation processes based on well-defined notions of algorithmic fairness
  • Score-based weighted sampling to preserve minority subgroup densities
  • Techniques to break spurious correlations while maintaining data utility

2. Fair Latent Representations

We propose learning syntax-agnostic, model-agnostic fair latent representations that separate fairness optimization from data generation:

  • Enables more stable training by decoupling fairness constraints from generator optimization
  • Reduces computational demands by generating data in low-dimensional spaces
  • Provides flexibility to use different generative model architectures (GANs, Diffusion Models, VAEs)

3. Intersectional Fairness

We address the challenge of intersectional bias, where multiple sensitive attributes interact to create compounded discrimination:

  • Novel fairness metrics that capture intersectional disparities
  • Generation techniques that ensure fairness across intersecting demographic groups
  • Applications to temporal and sequential data domains

Key Contributions

Our work has resulted in:

  • Novel architectures: Bt-GAN, FLDGM, and Fair distillation frameworks that achieve state-of-the-art performance in fair synthetic data generation and fair representation learning
  • Empirical validation: Extensive experiments on real-world healthcare datasets (MIMIC-III, IV) demonstrating improved fairness without sacrificing utility
  • Open-source tools: Publicly available implementations enabling reproducible research and practical applications

Impact and Applications

This research enables:

  • Fair healthcare AI: Training clinical decision support systems that provide equitable care across patient demographics
  • Bias mitigation: Identifying and correcting biases in existing datasets and models

Technical Approach

Our methods employ:

  • Deep generative models: GANs, Diffusion Models, Variational Autoencoders
  • Fairness metrics: Demographic parity, equalized odds, individual fairness measures
  • Optimization techniques: Adversarial training, score-based sampling, information constraints
  • Evaluation frameworks: Comprehensive assessment of utility (quality, diversity, fidelity) and fairness trade-offs

Related Publications

FairRep: Mitigating Intersectional Bias through Fair Representation Learning

Md Fahim Sikder, Daniel de Leng, Fredrik Heintz

2026

Representative Synthetic Data for Fair Decision Making

Md Fahim Sikder

2025

Deep generative models and representation learning techniques have become essential to modern machine learning, enabling both the creation of synthetic data and the extraction of meaningful features …

Promoting Intersectional Fairness through Knowledge Distillation

Md Fahim Sikder, Resmi Ramachandranpillai, Daniel de Leng, Fredrik Heintz

ECAI 2025 – 2025

As Artificial Intelligence-driven decision-making systems become increasingly popular, ensuring fairness in their outcomes has emerged as a critical and urgent challenge. AI models, often trained on …

FairX: A comprehensive benchmarking tool for model analysis using fairness, utility, and eXplainability

Md Fahim Sikder, Resmi Ramachandranpillai, Daniel de Leng, Fredrik Heintz

AEQUITAS, co-located with ECAI 2024 – 2024

We present FairX, an open-source Python-based benchmarking tool designed for the comprehensive analysis of models under the umbrella of fairness, utility, and eXplainability (XAI). FairX enables users …

Bt-GAN: Generating Fair Synthetic Healthdata via Bias-transforming Generative Adversarial Networks

Resmi Ramachandranpillai, Md Fahim Sikder, David Bergström, Fredrik Heintz

JAIR – 2024

Synthetic data generation offers a promising solution to enhance the usefulness of Electronic Healthcare Records (EHR) by generating realistic de-identified data. However, the existing literature …

Generative Models for Time-Series and Fair Data Generation

Md Fahim Sikder, Resmi Ramachandranpillai, Daniel de Leng, Fredrik Heintz

SSBA/SSDL 2024 as extended abstract – 2024

Generating Private and Fair Long-Sequenced Longitudinal Healthcare Records

Md Fahim Sikder, Resmi Ramachandranpillai, Fredrik Heintz

SAIS 2023 as extended abstract – 2023

Generating long-sequenced longitudinal healthcare records is critical as it has numerous potential applications. Long-sequenced longitudinal data allow us to better understand and find patterns from …

Fair Latent Deep Generative Models (FLDGM) for Syntax-agnostic and Fair Synthetic Data Generation

Resmi Ramachandranpillai, Md Fahim Sikder, Fredrik Heintz

ECAI 2023 – 2023

Deep Generative Models (DGMs) for generating synthetic data with properties such as quality, diversity, fidelity, and privacy is an important research topic. Fairness is one particular aspect that has …

Measuring the Fairness, Privacy and Accuracy of Time-Series Generative Models - A Survey

David Bergström, Md Fahim Sikder, Sania Partovian, Resmi Ramachandranpillai, Mattias Tiger, Fredrik Heintz

2021