Farayola, Michael Mayowa (2025) Fairness of Artificial Intelligence in Predicting Recidivism Risk and Beyond. PhD thesis, Dublin City University.
Abstract
This thesis designs, evaluates, and deploys fairness-aware artificial intelligence (AI) systems for recidivism prediction and other high-stakes domains. It addresses persistent gaps in the literature by integrating fairness interventions across the AI lifecycle, improving fairness–accuracy trade-offs, and developing methods that expose intersectional harms. The research advances multi-phase fairness pipelines that integrate pre-, in-, and post-processing techniques, supported by optimizationinformed evaluation and intersectional auditing frameworks. Grounded in trustworthy AI and socio-technical governance principles, the study
aligns fairness interventions with ethical and procedural standards for responsible AI in criminal justice and beyond. Empirical validation uses four datasets: COMPAS, RisCanvi (a curated dataset), Adult Income, and an Irish health insurance dataset. Standardized protocols and fairness metrics, Statistical Parity Difference, Disparate Impact, Equal Opportunity Difference, and Predictive Equality Difference, guide analysis and model evaluation. The thesis extends the AIF360 toolkit through enhanced algorithms (DIR+,AD+) capable of handling multi-valued protected attributes and improving training stability. Results show that integrated pipelines outperform isolated methods, achieving consistent fairness gains with limited accuracy reduction. Fairness-aware oversampling mitigates subgroup imbalance but requires careful validation to avoid synthetic bias. Intersectional auditing reveals disparities masked in aggregate measures and highlights the contextual nature of fairness.
Conceptually, the thesis contributes the Justice-Aware AI Fairness (JAAF) framework, linking technical fairness methods with intersectional and governance-based accountability. Overall, it delivers integrated fairness pipelines, optimization-based evaluation tools, and guidelines for fair, transparent, and trustworthy AI systems in high-stakes decision-making.
Metadata
| Item Type: | Thesis (PhD) |
|---|---|
| Date of Award: | 30 December 2025 |
| Refereed: | No |
| Supervisor(s): | Tal, Irina, Connolly, Regina and Bendechache, Malika |
| Subjects: | Computer Science > Artificial intelligence Computer Science > Machine learning |
| DCU Faculties and Centres: | DCU Faculties and Schools > Faculty of Engineering and Computing DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing |
| Use License: | This item is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 License. View License |
| Funders: | LERO |
| ID Code: | 32106 |
| Deposited On: | 14 Apr 2026 11:27 by Irina Tal . Last Modified 14 Apr 2026 11:27 |
Documents
Full text available as:
Preview |
PDF
- Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Creative Commons: Attribution-Noncommercial-No Derivative Works 4.0 2MB |
Downloads
Downloads
Downloads per month over past year
Archive Staff Only: edit this record