Other publication

Detecting demographic bias in automatically generated personas




AuthorsSalminen J., Soongyo J., Jansen B.

Conference nameConference on Human Factors in Computing Systems

PublisherAssociation for Computing Machinery

Publication year2019

Book title CHI EA '19 Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems

Journal name in sourceConference on Human Factors in Computing Systems - Proceedings

Number of pages6

ISBN978-1-4503-5971-9

DOIhttps://doi.org/10.1145/3290607.3313034


Abstract

We investigate the existence of demographic bias in automatically
generated personas by producing personas from YouTube Analytics data.
Despite the intended objectivity of the methodology, we find elements of
bias in the data-driven personas. The bias is highest when doing an
exact match comparison, and the bias decreases when comparing at age or
gender level. The bias also decreases when increasing the number of
generated personas. For example, the smaller number of personas resulted
in underrepresentation of female personas. This suggests that a higher
number of personas gives a more balanced representation of the user
population and a smaller number increases biases. Researchers and
practitioners developing data-driven personas should consider the
possibility of algorithmic bias, even unintentional, in their personas
by comparing the personas against the underlying raw data.



Last updated on 2024-26-11 at 13:40