Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 41 tok/s
Gemini 2.5 Pro 46 tok/s Pro
GPT-5 Medium 21 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 91 tok/s Pro
Kimi K2 178 tok/s Pro
GPT OSS 120B 474 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Illuminating the Black Box: A Psychometric Investigation into the Multifaceted Nature of Large Language Models (2312.14202v1)

Published 21 Dec 2023 in cs.CL and cs.AI

Abstract: This study explores the idea of AI Personality or AInality suggesting that LLMs exhibit patterns similar to human personalities. Assuming that LLMs share these patterns with humans, we investigate using human-centered psychometric tests such as the Myers-Briggs Type Indicator (MBTI), Big Five Inventory (BFI), and Short Dark Triad (SD3) to identify and confirm LLM personality types. By introducing role-play prompts, we demonstrate the adaptability of LLMs, showing their ability to switch dynamically between different personality types. Using projective tests, such as the Washington University Sentence Completion Test (WUSCT), we uncover hidden aspects of LLM personalities that are not easily accessible through direct questioning. Projective tests allowed for a deep exploration of LLMs cognitive processes and thought patterns and gave us a multidimensional view of AInality. Our machine learning analysis revealed that LLMs exhibit distinct AInality traits and manifest diverse personality types, demonstrating dynamic shifts in response to external instructions. This study pioneers the application of projective tests on LLMs, shedding light on their diverse and adaptable AInality traits.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. ChatGPT Can Accurately Predict Public Figures’ Perceived Personalities Without Any Training.
  2. Dair.AI. 2023. Prompt Engineering Guide.
  3. Do personality tests generalize to Large Language Models? arXiv:2311.05297.
  4. Edwards, B. 2023. AI-powered Bing Chat spills its secrets via prompt injection attack.
  5. Systematic Evaluation of GPT-3 for Zero-Shot Personality Estimation. arXiv:2306.01183.
  6. Evaluating and Inducing Personality in Pre-trained Language Models.
  7. PersonaLLM: Investigating the Ability of Large Language Models to Express Big Five Personality Traits. arXiv:2305.02547.
  8. Introducing the Short Dark Triad (SD3): A Brief Measure of Dark Personality Traits. Assessment, 21(1): 28–41. PMID: 24322012.
  9. Does GPT-3 Demonstrate Psychopathy? Evaluating Large Language Models from a Psychological Perspective. arXiv:2212.10529.
  10. Loevinger, J.; et al. 2014. Measuring ego development. Psychology Press.
  11. Editing Personality for LLMs. arXiv:2310.02168.
  12. The Myers-Briggs Type Indicator: Manual (1962). Consulting Psychologists Press.
  13. OSPP. 2019. Open-Source Psychometric Project(OSPP).
  14. Do LLMs Possess a Personality? Making the MBTI Test an Amazing Evaluation for Large Language Models. arXiv:2307.16180.
  15. Paoli, S. D. 2023. Improved prompting and process for writing user personas with LLMs, using qualitative interviews: Capturing behaviour and personality traits of users. arXiv:2310.06391.
  16. Sigmund, F. 1997. The Interpretation of Dreams. Wordsworth Editions.
  17. Cue-CoT: Chain-of-thought Prompting for Responding to In-depth Dialogue Questions with LLMs. arXiv:2305.11792.
  18. Does Role-Playing Chatbots Capture the Character Personalities? Assessing Personality Traits for Role-Playing Chatbots. arXiv:2310.17976.
  19. Wilson, E. W. 2023. The Sentence Completion Test.
Citations (2)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces AInality as measurable, human-like traits in LLMs through a rigorous psychometric and machine learning framework.
  • It employs tests like MBTI, BFI, SD3, and WUSCT alongside advanced prompt engineering to reveal consistent and identifiable personality profiles.
  • The study demonstrates that LLMs exhibit adaptable personalities, suggesting opportunities for customized user-AI interactions via prompt modulation.

Exploring AInality in LLMs Through Psychometric Testing

This paper introduces the concept of "AInality" to describe the human-like personality traits exhibited by LLMs. The authors explore whether LLMs possess consistent and identifiable personality traits that can be measured using traditional psychometric tests (2312.14202). The paper employs a combination of prompt engineering and machine learning techniques to investigate the AInality of LLMs, examining its consistency, discernibility, and potential for modulation.

Methodology

The research methodology is structured around the application of psychometric tests, prompt engineering techniques, and machine learning models to analyze the personality traits of LLMs. The paper uses four psychometric tests: the Myers-Briggs Type Indicator (MBTI), the Big Five Inventory (BFI), the Short Dark Triad (SD3), and the Washington University Sentence Completion Test (WUSCT). Prompt engineering techniques, including few-shot prompting, chain-of-thought (CoT) prompting, role-play prompting, and generated knowledge prompting, are used to elicit responses from LLMs. The collected data is then analyzed using machine learning models such as Random Forest, Logistic Regression, SVM, Naive Bayes, and Neural Networks.

Experimental Results and Analysis

The paper's initial experiments aimed to determine whether LLMs exhibit personality traits. The authors found notable variations in the MBTI test results for each user, indicating distinct AInalities exhibited by the LLMs. The BFI test results for ChatGPT and Bard showed mixed alignment with their MBTI types, suggesting that LLM AInality may be more complex than either test can fully capture. Figure 1

Figure 1: ChatGPT's Big Five Inventory test results, illustrating scores for Extraversion, Emotional Stability, Agreeableness, Conscientiousness, and Intellect.

The SD3 test results indicated that ChatGPT had relatively low scores for Machiavellianism and Psychopathy but a moderate score for Narcissism, while Bard had moderate scores for Machiavellianism and Narcissism and a low score for Psychopathy.

To assess the consistency of AInality, the researchers employed role-play techniques, instructing LLMs to adopt specific MBTI personality types and respond to questionnaires accordingly. LLMs demonstrated a strong understanding of each of the sixteen personality types and the capability to choose reactions in line with the specified personality type.

To identify AInality, the authors gathered MBTI test results from Bard and ChatGPT and utilized machine learning models for classification. The models, including Random Forest, Logistic Regression, and SVM, achieved high accuracy in predicting the LLMs, suggesting that LLMs exhibit consistent AInality traits, even when engaging in role-play scenarios that involve assuming different personalities.

The Washington University Sentence Completion Test (WUSCT) was used to uncover deeper aspects of LLM personality traits. The analysis of the WUSCT responses allowed the researchers to gain insights into the emotional, psychological, and AInality aspects of LLMs, revealing layers that might not be apparent through conventional self-reporting tests.

Implications and Future Directions

The paper's findings have several implications for the development and application of LLMs. The authors demonstrate that LLMs can exhibit diverse and adaptable AInality traits, which can be modulated through prompt engineering. This opens up possibilities for customizing LLMs according to users' unique personality types, enhancing user experiences, and improving human-AI interactions.

The paper also highlights the potential of using projective tests, such as the WUSCT, to gain deeper insights into the AInality structures of LLMs. The authors suggest that future research should focus on developing new AInality psychometric tests specifically tailored for LLMs, targeting their persona characteristics. This could lead to a more comprehensive understanding of LLM behavior and cognition, as well as the development of more user-centric AI systems.

Conclusion

This paper provides a comprehensive psychometric investigation into the multifaceted nature of LLMs, introducing the concept of AInality and demonstrating its discernibility, consistency, characterization, and malleability through prompt engineering (2312.14202). The use of machine learning analysis and the application of projective tests such as the WUSCT represent a significant contribution to the understanding of LLM psychological traits and the potential for creating more personalized and effective AI interactions.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube