Please use this identifier to cite or link to this item: http://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/9816
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorBorah, Abhinash-
dc.contributor.authorH, SHASHWATI-
dc.date.accessioned2025-05-13T04:35:08Z-
dc.date.available2025-05-13T04:35:08Z-
dc.date.issued2025-05-
dc.identifier.citation168en_US
dc.identifier.urihttp://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/9816-
dc.description.abstractExperiments in economics and social psychology have repeatedly shown that humans are not solely self-interested but also other-regarding. Models of human behavior have been developed to explain the diverse social preferences exhibited by individuals, such as fairness, trust, and reciprocity, in both individualistic and group settings. Recently, another such model has been revolutionizing the social sciences—Large Language Models (LLMs), deep-learning models that are increasingly being used to simulate human decision-making and replicate strategic behavior. In this study, we investigate the social preferences of GPT-4o-mini, a potential computational model of human behavior, using economic games such as the Other-Other Task, Dictator, Ultimatum, and Trust Games. While many studies have examined LLM behavior in individualistic settings, we extend this analysis to group contexts inspired by the Minimal Group Paradigm and Natural Identities frameworks. Specifically, we explore whether LLMs exhibit human-like biases such as ingroup favoritism in decision-making. Additionally, we assess the model’s distributional preferences and reciprocity concerns by replicating the experiments of Chen & Li (2009). Our findings reveal several key insights into the model’s behavior across economic games. In games with minimal groups, self-interest in the dictator game reduces the discriminatory tendencies observed in the other-other task. Fairness emerges as a dominant concern, as evidenced by the model’s reasoning and higher offers in dictator games. Ingroup bias is observed at certain allocation levels among ultimatum game responders and trust game trustees. However, in experiments involving natural identities of religion (Christian, Hindu, and Muslim) and gender (Female and Male), bias appears only among trustees in the trust game. Additionally, in the Chen & Li (2009) games, the model struggles to translate words into actions and to grasp the intentions of others and the consequences of its own decisions. Overall, this study underscores the potential of LLMs as valuable tools for advancing social science research while also highlighting their limitations in capturing the complexities of human decision-making.en_US
dc.language.isoenen_US
dc.subjectLarge Language Modelsen_US
dc.subjectDiscriminationen_US
dc.subjectMinimal Group Paradigmen_US
dc.subjectNatural Identitiesen_US
dc.subjectSocial Preferencesen_US
dc.subjectGPTen_US
dc.subjectArtificial Intelligenceen_US
dc.subjectOther-Other Tasken_US
dc.subjectDictator Gameen_US
dc.subjectUltimatum Gameen_US
dc.subjectTrust Gameen_US
dc.titleUnderstanding Discrimination in Large Language Modelsen_US
dc.typeThesisen_US
dc.description.embargoOne Yearen_US
dc.type.degreeBS-MSen_US
dc.contributor.departmentDept. of Humanities and Social Sciencesen_US
dc.contributor.registration20201090en_US
Appears in Collections:MS THESES

Files in This Item:
File Description SizeFormat 
20201090_Shashwati_H_MS_Thesis.pdfMS Thesis11.75 MBAdobe PDFView/Open    Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.