Mitigating Bias in Recommender Systems

Question: Consider how to counter the radicalizing effects of recommender systems or ways to prevent algorithmic discrimination.

As we know, recommender systems are used to make decisions that may affect users such as job applications, getting a loan, movie selection and the like. Such decisions can affect human rights and undermine the public trust. If a user’s opinion is negative, it is likely to thwart the development of machine learning recommender systems and its positive social and economic potential. Also, users of different ages or genders may not obtain similar utility from the system, especially if their group is a relatively small subset of the user base. On the other hand, when designed well, machine learning systems can help mitgiate the type of human bias in the decision making.

Ways to consider the prevention of algorithmic discrimination include:

Identifying and eliminating bias or discrimination that can result from machine learning applications is not an easy task. However, developers can work together with relevant stakeholders to leverage machine learning in a way that includes and benefits people, and prevents discrimination.