Personality Shapes Gender Bias in Persona-Conditioned LLM Narratives Across English and Hindi: An Empirical Investigation
Abstract
Persona-conditioned large language models exhibit context-dependent gender bias that varies with personality trait frameworks and across languages.
Large Language Models (LLMs) are increasingly deployed in persona-driven applications such as education, customer service, and social platforms, where models are prompted to adopt specific personas when interacting with users. While persona conditioning can improve user experience and engagement, it also raises concerns about how personality cues may interact with gender biases and stereotypes. In this work, we present a controlled study of persona-conditioned story generation in English and Hindi, where each story portrays a working professional in India producing context-specific artifacts (e.g., lesson plans, reports, letters) under systematically varied persona gender, occupational role, and personality traits from the HEXACO and Dark Triad frameworks. Across 23,400 generated stories from six state-of-the-art LLMs, we find that personality traits are significantly associated with both the magnitude and direction of gender bias. In particular, Dark Triad personality traits are consistently associated with higher gender-stereotypical representations compared to socially desirable HEXACO traits, though these associations vary across models and languages. Our findings demonstrate that gender bias in LLMs is not static but context-dependent. This suggests that persona-conditioned systems used in real-world applications may introduce uneven representational harms, reinforcing gender stereotypes in generated educational, professional, or social content.
Community
This paper introduces a multilingual, persona-conditioned bias evaluation framework showing that personality traits act as systematic modulators of gender bias in LLM narrative generation, with Dark Triad traits amplifying stereotypes, HEXACO traits partially attenuating them, and these effects often exceeding the influence of explicit gender labels.
โก๏ธ ๐๐๐ฒ ๐๐ข๐ ๐ก๐ฅ๐ข๐ ๐ก๐ญ๐ฌ ๐จ๐ ๐๐๐ซ๐ฌ๐จ๐ง๐๐ฅ๐ข๐ญ๐ฒ-๐๐จ๐ง๐๐ข๐ญ๐ข๐จ๐ง๐๐ ๐๐ข๐๐ฌ ๐๐จ๐๐ฎ๐ฅ๐๐ญ๐ข๐จ๐ง ๐ ๐ซ๐๐ฆ๐๐ฐ๐จ๐ซ๐ค:
๐งช ๐ท๐๐๐๐๐๐๐๐๐๐ ๐๐ ๐ ๐ญ๐๐๐๐๐๐๐-๐ช๐๐๐๐๐๐๐ ๐ช๐๐๐๐๐๐ ๐ฝ๐๐๐๐๐๐๐:
Introduces a controlled 23,400-artifact multilingual benchmark spanning 6 model families (LLM, MoE, SSM, LRM, SLM), 50 occupations, 9 personality traits ร 2 levels, 3 gender conditions, and English/Hindi, enabling systematic measurement of personalityโgender interaction effects in generation bias. Novel finding: gender bias is conditional, not staticโDark Triad traits (especially Machiavellianism, Psychopathy) amplify stereotypical outputs, while Openness/Emotionality attenuate them.
๐งฉ ๐ช๐๐๐๐๐๐๐
-๐ฉ๐๐๐๐
๐บ๐๐๐๐๐๐๐ ๐ฉ๐๐๐ ๐ด๐๐๐๐๐ ๐๐๐ ๐ท๐๐๐๐๐๐-๐ช๐๐๐
๐๐๐๐๐๐๐
๐ต๐๐๐๐๐๐๐๐๐:
Proposes a sentence-level stereotype centroid scoring framework using multilingual SBERT embeddings and a difference-of-cosines bias score against male/female stereotype centroids, aggregated via maximum salience over narrative sentences. This moves beyond benchmark-style classification by localizing where stereotypes emerge inside generated artifacts. Human validation (ฮบโ0.66โ0.69) supports the metricโs alignment with perceived stereotyping.
๐ง ๐ด๐๐๐๐๐๐๐๐๐๐๐ ๐ท๐๐๐๐๐๐๐๐๐๐โ๐ฎ๐๐๐
๐๐ ๐ฐ๐๐๐๐๐๐๐๐๐๐ ๐จ๐๐๐๐๐๐๐:
Shows that personality coefficients can exceed explicit gender effects, reframing prompt persona design as an alignment problem rather than prompt styling. Reveals cross-linguistic asymmetry: Hindi exhibits stronger baseline male-stereotypical skew, while English shows stronger personality-driven modulation. Importantly, the directional pattern generalizes across architectures, suggesting persona-induced bias amplification may be a broad property of LLMs rather than model-specific behavior.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Invisible Influences: Investigating Implicit Intersectional Biases through Persona Engineering in Large Language Models (2026)
- Dual-Metric Evaluation of Social Bias in Large Language Models: Evidence from an Underrepresented Nepali Cultural Context (2026)
- Widespread Gender and Pronoun Bias in Moral Judgments Across LLMs (2026)
- The Binding Effect: Analyzing How Multi-Dimensional Cues Form Gender Bias in Instruction TTS (2026)
- ProText: A benchmark dataset for measuring (mis)gendering in long-form texts (2026)
- Gender Bias in MT for a Genderless Language: New Benchmarks for Basque (2026)
- Small Changes, Big Impact: Demographic Bias in LLM-Based Hiring Through Subtle Sociocultural Markers in Anonymised Resumes (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.23600 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper