TU Berlin

Efficient Algorithms (ALGO)Abstract Meike Zehlike

Page Content

to Navigation

Talk by Meike Zehlike

Discrimination Mitigation in Ranked Search Results

by Meike Zehlike

Personalization and algorithmic decision-making are essential tools in our daily lives. They decide about the media we consume, the stories we read, the people we meet, the places we visit, whether we get a job, or if our loan request is approved. It is therefore of societal and ethical importance to ask whether these algorithms eventually produce results that demote, marginalize, or exclude individuals belonging to an unprivileged group or a minority. Early work in the field of “Fairness, Accountability and Transparency in socio-technical Systems” shows that various forms of bias and discrimination can arise in these systems, leading to systematical disadvantages for certain individuals and societal groups, as well as into distortion of competition. However as Cathy O'Neill illustrates in various examples, biases in computer systems can be difficult to identify due to the system’s complexity, yet “[. . .] a biased system has the potential of a widespread impact. If the system  becomes a standard in the field the bias becomes pervasive [. . .]” and unlike in case of a biased individual, a biased system offers no possibility of negotiation to the victim.

In this talk i will present two methods to mitigate bias in ranked outputs: The first one reranks a given search result subject to fairness criteria that are based on a statistical significance test, while maintaining the ranking utility as high as possible. The second one modifies the loss function of a learning to rank method such that it considers not only the loss wrt to the training data, but also the amount of bias in a predicted result measured in terms of disparate exposure across different social groups.


at 15.00
Meike Zehlike

Back to the research colloquium site.

To top


Quick Access

Schnellnavigation zur Seite über Nummerneingabe