Papers
arxiv:2408.11247

Unboxing Occupational Bias: Grounded Debiasing LLMs with U.S. Labor Data

Published on Aug 20
ยท Submitted by amanchadha on Aug 22
Authors:
,

Abstract

Large Language Models (LLMs) are prone to inheriting and amplifying societal biases embedded within their training data, potentially reinforcing harmful stereotypes related to gender, occupation, and other sensitive categories. This issue becomes particularly problematic as biased LLMs can have far-reaching consequences, leading to unfair practices and exacerbating social inequalities across various domains, such as recruitment, online content moderation, or even the criminal justice system. Although prior research has focused on detecting bias in LLMs using specialized datasets designed to highlight intrinsic biases, there has been a notable lack of investigation into how these findings correlate with authoritative datasets, such as those from the U.S. National Bureau of Labor Statistics (NBLS). To address this gap, we conduct empirical research that evaluates LLMs in a ``bias-out-of-the-box" setting, analyzing how the generated outputs compare with the distributions found in NBLS data. Furthermore, we propose a straightforward yet effective debiasing mechanism that directly incorporates NBLS instances to mitigate bias within LLMs. Our study spans seven different LLMs, including instructable, base, and mixture-of-expert models, and reveals significant levels of bias that are often overlooked by existing bias detection techniques. Importantly, our debiasing method, which does not rely on external datasets, demonstrates a substantial reduction in bias scores, highlighting the efficacy of our approach in creating fairer and more reliable LLMs.

Community

Paper author Paper submitter

Screenshot 2024-08-22 at 4.09.34โ€ฏPM.jpg

๐Ÿ“ Announcing our paper published at AAAI Spring Symposium 2024 that introduces a novel grounded debiasing approach for LLMs by incorporating U.S. labor data, effectively reducing occupational biases without relying on external datasets

  • ๐‚๐จ๐ฆ๐ฉ๐ซ๐ž๐ก๐ž๐ง๐ฌ๐ข๐ฏ๐ž ๐๐ข๐š๐ฌ ๐€๐ง๐š๐ฅ๐ฒ๐ฌ๐ข๐ฌ: The paper presents a robust framework for analyzing occupational biases in LLMs using authoritative U.S. labor data, revealing significant biases in existing models.
  • ๐„๐Ÿ๐Ÿ๐ž๐œ๐ญ๐ข๐ฏ๐ž ๐ƒ๐ž๐›๐ข๐š๐ฌ๐ข๐ง๐  ๐Œ๐ž๐ญ๐ก๐จ๐: The authors propose a simple debiasing mechanism that utilizes contextual examples from labor data, achieving up to a 65% reduction in bias across multiple LLMs.
  • ๐„๐ฆ๐ฉ๐ข๐ซ๐ข๐œ๐š๐ฅ ๐‘๐ž๐ฌ๐ฎ๐ฅ๐ญ๐ฌ: The study shows that the proposed method substantially outperforms existing techniques, highlighting the importance of grounded knowledge in creating fairer LLM outputs.
ยท

Hi @amanchadha ,

Congrats on this work! Are you planning to share any artifacts (datasets, models, a demo as a Space) on the hub?

Happy to assist :)

Cheers,
Niels from HF

Paper author Paper submitter

Published at AAAI Spring Symposium 2024.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2408.11247 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2408.11247 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2408.11247 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.