Abstract

Algorithmic bias refers to systematic and repeatable errors within a computer system that generate unfair outcomes, often favoring one group of users over others. Bias can enter through historically skewed training data, self-reinforcing feedback loops that perpetuate prior patterns, or design choices made during the development of the algorithm itself. Existing literature on algorithmic bias has discussed aspects such as how bias materializes in algorithmic outputs (Obermeyer et al., 2019) and the underlying epistemic opacity of the algorithms hindering experts’ capacity to trust a judgment (Lebovitz et al., 2022). What if bias is not an outcome of inattention to vulnerable populations or opacity of algorithms, but emanates from the specific attempts to help the vulnerable populations and to address the opacity of algorithms? Recently, limited research offers some evidence of this possibility of bias backfiring (Yan et al., 2024). Such impact can undermine trust in AI (Artificial Intelligence) systems, especially when fairness claims are made without transparency about trade-offs or unintended consequences. In order to explore this phenomenon, I intend to undertake an empirical study within a financial institution that actively engages in the development and use of algorithmic systems. The setting is relevant since financial services are high-stakes contexts, where algorithmic decisions can significantly impact individuals’ access to credit, insurance, and other financial products. The study will seek to answer two research questions: First, what specific guardrails and oversight mechanisms does the organization implement to assess the potential consequences of a bias mitigation strategy before its deployment. Second, once implemented, how does the organization evaluate the effectiveness of these checks in practice, and what mechanisms are employed to ensure that such interventions do not generate adverse outcomes over time. By examining these questions, the study aims to contribute to the growing body of research on fairness in algorithms, by providing an empirical account of how practitioners navigate the risk of bias-mitigation measures inadvertently backfiring.

Comments

tpp1309

Share

COinS