About

AI techniques based on big data and algorithmic processing are increasingly used to guide decisions in important societal spheres, including hiring decisions, university admissions, loan granting, and crime prediction. They are applied by search engines, Internet recommendation systems and social media bots, influencing our perceptions of political developments and even of scientific findings. However, there are growing concerns with regard to the epistemic and normative quality of AI evaluations and predictions. In particular, there is strong evidence that algorithms may sometimes amplify rather than eliminate existing bias and discrimination, and thereby have negative effects on social cohesion and on democratic institutions.

Scholarly reflection of these issues has begun but is still in its early stages and a lot of work remains to be done. In particular, we still lack a comprehensive understanding of how pertinent concepts of bias or discrimination should be interpreted in the context of AI and which technical options to combat bias and discrimination are both realistically possible and normatively justified. The research group “BIAS” examines these issues in an integrated, interdisciplinary project bringing together experts from philosophy, law, and computer science. Our shared research question is: How can standards of unbiased attitudes and non-discriminatory practices be met in big data analysis and algorithm-based decision-making? In approaching this question, we provide philosophical analyses of the relevant concepts and principles in the context of AI (“bias”, “discrimination”, “fairness”), investigate their adequate reception in pertinent legal frameworks (data protection, consumer, competition, anti-discrimination law), and develop concrete technical solutions (debiasing strategies, discrimination detection procedures etc.).