FairLoop: Software Support for Human-Centric Fairness in Predictive Business Process Monitoring

Möhrlein, Felix, Käppel, Martin, Neuberger, Julian, Weinzierl, Sven, Ackermann, Lars, Matzner, Martin, Jablonski, Stefan

arXiv.org Artificial Intelligence 

Sensitive attributes like gender or age can lead to unfair predictions in machine learning tasks such as predictive business process monitoring, particularly when used without considering context. We present FairLoop1, a tool for human-guided bias mitigation in neural network-based prediction models. FairLoop distills decision trees from neural networks, allowing users to inspect and modify unfair decision logic, which is then used to fine-tune the original model towards fairer predictions. Compared to other approaches to fairness, FairLoop enables context-aware bias removal through human involvement, addressing the influence of sensitive attributes selectively rather than excluding them uniformly.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found