Probing into the Fairness of Large Language Models: A Case Study of ChatGPT

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Understanding and addressing unfairness in LLMs are crucial for responsible AI deployment. However, there is a limited number of quantitative analyses and in-depth studies regarding fairness evaluations in LLMs, especially when applying LLMs to high-stakes fields. This work aims to fill this gap by providing a systematic evaluation of the effectiveness and fairness of LLMs using ChatGPT as a study case. We focus on assessing ChatGPT's performance in high-takes fields including education, criminology, finance and healthcare. To conduct a thorough evaluation, we consider both group fairness and individual fairness metrics. We also observe the disparities in ChatGPT's outputs under a set of biased or unbiased prompts. This work contributes to a deeper understanding of LLMs' fairness performance, facilitates bias mitigation and fosters the development of responsible AI systems. Code and data are open-sourced on GitHub.

Original languageEnglish (US)
Title of host publication2024 58th Annual Conference on Information Sciences and Systems, CISS 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350369298
DOIs
StatePublished - 2024
Externally publishedYes
Event58th Annual Conference on Information Sciences and Systems, CISS 2024 - Princeton, United States
Duration: Mar 13 2024Mar 15 2024

Publication series

Name2024 58th Annual Conference on Information Sciences and Systems, CISS 2024

Conference

Conference58th Annual Conference on Information Sciences and Systems, CISS 2024
Country/TerritoryUnited States
CityPrinceton
Period3/13/243/15/24

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Information Systems
  • Safety, Risk, Reliability and Quality
  • Control and Optimization
  • Modeling and Simulation
  • Computational Theory and Mathematics

Keywords

  • ChatGPT
  • Fairness
  • Large Language Model

Cite this