Bias mitigation with AIF360: A comparative study
Chapter
Published version
Åpne
Permanent lenke
https://hdl.handle.net/11250/2764230Utgivelsesdato
2020Metadata
Vis full innførselSamlinger
Originalversjon
NIK Norsk informatikkonferanse. 2020, 1.Sammendrag
The use of artificial intelligence for decision making raises concerns about the societal impact of such systems. Traditionally, the product of a human decision-maker are governed by laws and human values. Decision-making is now being guided - or in some cases, replaced by machine learning classification which may reinforce and introduce bias. Algorithmic bias mitigation is explored as an approach to avoid this, however it does come at a cost: efficiency and accuracy. We conduct an empirical analysis of two off-the-shelf bias mitigation techniques from the AIF360 toolkit on a binary classification task. Our preliminary results indicate that bias mitigation is a feasible approach to ensuring group fairness.