Digital Repository

Understanding Discrimination in Large Language Models

Show simple item record

dc.contributor.advisor Borah, Abhinash
dc.contributor.author H, SHASHWATI
dc.date.accessioned 2025-05-13T04:35:08Z
dc.date.available 2025-05-13T04:35:08Z
dc.date.issued 2025-05
dc.identifier.citation 168 en_US
dc.identifier.uri http://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/9816
dc.description.abstract Experiments in economics and social psychology have repeatedly shown that humans are not solely self-interested but also other-regarding. Models of human behavior have been developed to explain the diverse social preferences exhibited by individuals, such as fairness, trust, and reciprocity, in both individualistic and group settings. Recently, another such model has been revolutionizing the social sciences—Large Language Models (LLMs), deep-learning models that are increasingly being used to simulate human decision-making and replicate strategic behavior. In this study, we investigate the social preferences of GPT-4o-mini, a potential computational model of human behavior, using economic games such as the Other-Other Task, Dictator, Ultimatum, and Trust Games. While many studies have examined LLM behavior in individualistic settings, we extend this analysis to group contexts inspired by the Minimal Group Paradigm and Natural Identities frameworks. Specifically, we explore whether LLMs exhibit human-like biases such as ingroup favoritism in decision-making. Additionally, we assess the model’s distributional preferences and reciprocity concerns by replicating the experiments of Chen & Li (2009). Our findings reveal several key insights into the model’s behavior across economic games. In games with minimal groups, self-interest in the dictator game reduces the discriminatory tendencies observed in the other-other task. Fairness emerges as a dominant concern, as evidenced by the model’s reasoning and higher offers in dictator games. Ingroup bias is observed at certain allocation levels among ultimatum game responders and trust game trustees. However, in experiments involving natural identities of religion (Christian, Hindu, and Muslim) and gender (Female and Male), bias appears only among trustees in the trust game. Additionally, in the Chen & Li (2009) games, the model struggles to translate words into actions and to grasp the intentions of others and the consequences of its own decisions. Overall, this study underscores the potential of LLMs as valuable tools for advancing social science research while also highlighting their limitations in capturing the complexities of human decision-making. en_US
dc.language.iso en en_US
dc.subject Large Language Models en_US
dc.subject Discrimination en_US
dc.subject Minimal Group Paradigm en_US
dc.subject Natural Identities en_US
dc.subject Social Preferences en_US
dc.subject GPT en_US
dc.subject Artificial Intelligence en_US
dc.subject Other-Other Task en_US
dc.subject Dictator Game en_US
dc.subject Ultimatum Game en_US
dc.subject Trust Game en_US
dc.title Understanding Discrimination in Large Language Models en_US
dc.type Thesis en_US
dc.description.embargo One Year en_US
dc.type.degree BS-MS en_US
dc.contributor.department Dept. of Humanities and Social Sciences en_US
dc.contributor.registration 20201090 en_US


Files in this item

This item appears in the following Collection(s)

  • MS THESES [1970]
    Thesis submitted to IISER Pune in partial fulfilment of the requirements for the BS-MS Dual Degree Programme/MSc. Programme/MS-Exit Programme

Show simple item record

Search Repository


Advanced Search

Browse

My Account