Skip to content

DerekEubanks/Algorithmic-Fairness-Analysis-in-Employee-Promotion-Models

Repository files navigation

⚖️ Algorithmic-Fairness-Analysis-in-Employee-Promotion-Models

📌 Overview

This project investigates fairness and bias in machine learning models used for employee promotion decisions. Using a real-world HR dataset, the goal is to evaluate whether predictive models produce equitable outcomes across demographic groups—and to explore methods for mitigating bias while maintaining model performance.

The work emphasizes a key challenge in modern AI systems: balancing predictive accuracy with fairness and ethical responsibility.

🎯 Objectives • Predict whether an employee is promoted based on historical data • Measure fairness using standard bias metrics • Identify disparities across demographic groups • Apply bias mitigation techniques and evaluate tradeoffs

📊 Dataset • ~54,000 employee records • Features include: • Education level • Years of service • Training scores • Previous performance • Department and region • Demographic attributes (used for fairness evaluation)

⚙️ Methodology

  1. Data Preprocessing • Cleaned missing and inconsistent values • Encoded categorical variables • Split data into training/testing sets

  2. Model Development • Implemented Logistic Regression as a baseline classifier • Trained model to predict promotion outcomes

  3. Fairness Evaluation

Measured bias using: • Statistical Parity Difference (SPD) → Difference in positive prediction rates between groups • Equal Opportunity Difference (EOD) → Difference in true positive rates between groups

These metrics quantify whether certain groups are systematically favored or disadvantaged.

🛠️ Bias Mitigation

Applied reweighing techniques to reduce bias in training data by adjusting sample importance across groups.

Evaluated how mitigation impacts: • Model accuracy • Precision and recall • Fairness metrics (SPD, EOD)

📈 Results • Baseline model achieved strong predictive performance • Detectable bias was present across demographic groups • Reweighing reduced bias (improved SPD/EOD) • However, fairness improvements came with tradeoffs in accuracy

👉 Key takeaway:

Improving fairness often reduces predictive performance—highlighting the need for careful, context-dependent decision-making in real-world systems.

🧠 Key Insights • Bias can persist even when sensitive attributes are not explicitly used • Fairness must be measured explicitly, not assumed • There is no single “best” model—only tradeoffs between competing objectives • Ethical AI requires both technical methods and human judgment

About

Analyzed algorithmic fairness in employee promotion outcomes using real HR data (~54k records). Implemented Statistical Parity Difference and Equal Opportunity Difference, applied reweighing for bias mitigation, and evaluated tradeoffs between fairness and model performance using Logistic Regression.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors