Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 49 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 19 tok/s Pro
GPT-5 High 16 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 172 tok/s Pro
GPT OSS 120B 472 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Identifying Multiple Personalities in Large Language Models with External Evaluation (2402.14805v1)

Published 22 Feb 2024 in cs.CL and cs.AI

Abstract: As LLMs are integrated with human daily applications rapidly, many societal and ethical concerns are raised regarding the behavior of LLMs. One of the ways to comprehend LLMs' behavior is to analyze their personalities. Many recent studies quantify LLMs' personalities using self-assessment tests that are created for humans. Yet many critiques question the applicability and reliability of these self-assessment tests when applied to LLMs. In this paper, we investigate LLM personalities using an alternate personality measurement method, which we refer to as the external evaluation method, where instead of prompting LLMs with multiple-choice questions in the Likert scale, we evaluate LLMs' personalities by analyzing their responses toward open-ended situational questions using an external machine learning model. We first fine-tuned a Llama2-7B model as the MBTI personality predictor that outperforms the state-of-the-art models as the tool to analyze LLMs' responses. Then, we prompt the LLMs with situational questions and ask them to generate Twitter posts and comments, respectively, in order to assess their personalities when playing two different roles. Using the external personality evaluation method, we identify that the obtained personality types for LLMs are significantly different when generating posts versus comments, whereas humans show a consistent personality profile in these two different situations. This shows that LLMs can exhibit different personalities based on different scenarios, thus highlighting a fundamental difference between personality in LLMs and humans. With our work, we call for a re-evaluation of personality definition and measurement in LLMs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. Musiclm: Generating music from text. arXiv preprint arXiv:2301.11325.
  2. American Psychological Association. 1983. Publications Manual. American Psychological Association, Washington, DC.
  3. American Psychological Association. 2023. Definition of Personality - https://www.apa.org/topics/personality.
  4. Personality testing of gpt-3: Limited temporal reliability, but highlighted social desirability of gpt-3’s personality instruments results. arXiv preprint arXiv:2306.04308.
  5. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  6. Graham Caron and Shashank Srivastava. 2022. Identifying and manipulating the personality traits of language models. arXiv preprint arXiv:2212.10276.
  7. Raymond B Cattell. 1943a. The description of personality: Basic traits resolved into clusters. The journal of abnormal and social psychology, 38(4):476.
  8. Raymond B Cattell. 1943b. The description of personality. i. foundations of trait measurement. Psychological review, 50(6):559.
  9. Unleashing the potential of prompt engineering in large language models: a comprehensive review. arXiv preprint arXiv:2310.14735.
  10. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  11. John M Digman. 1990. Personality structure: Emergence of the five-factor model. Annual review of psychology, 41(1):417–440.
  12. Katrina Faust. 2019. Myers-briggs type indicator (mbti) overview - https://leadx.org/articles/mbti-myers-briggs-type-indicator-overview/.
  13. Lewis R Goldberg. 1990. An alternative" description of personality": the big-five factor structure. Journal of personality and social psychology, 59(6):1216.
  14. Lewis R Goldberg. 1993. The structure of phenotypic personality traits. American psychologist, 48(1):26.
  15. Personality consistency and situational influences on behavior. Journal of Management, 45(8):3204–3234.
  16. Investigating the applicability of self-assessment tests for personality measurement of large language models. arXiv preprint arXiv:2309.08163.
  17. The curious case of neural text degeneration. arXiv preprint arXiv:1904.09751.
  18. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
  19. Chatgpt an enfj, bard an istj: Empirical study on personalities of large language models. arXiv preprint arXiv:2305.19926.
  20. Using a language model to generate music in its symbolic domain while controlling its perceived emotion. IEEE Access.
  21. Jaeho Jeon and Seongyong Lee. 2023. Large language models in education: A focus on the complementary relationship between human teachers and chatgpt. Education and Information Technologies, pages 1–20.
  22. Mpi: Evaluating and inducing personality in pre-trained language models. arXiv preprint arXiv:2206.07550.
  23. Evaluating and inducing personality in pre-trained language models.
  24. John A Johnson. 2014. Measuring thirty facets of the five factor model with a 120-item public domain inventory: Development of the ipip-neo-120. Journal of research in personality, 51:78–89.
  25. Estimating the personality of white-box language models. arXiv e-prints, pages arXiv–2204.
  26. Psy-llm: Scaling up global mental health psychological services with ai-based large language models. arXiv preprint arXiv:2307.11991.
  27. Yash Mehta. 2023. Personality prediction. https://github.com/yashsmehta/personality-prediction.git. GitHub repository.
  28. Bottom-up and top-down: Predicting personality with psycholinguistic and language model features. In 2020 IEEE International Conference on Data Mining (ICDM), pages 1184–1189. IEEE.
  29. Who is gpt-3? an exploration of personality, values and demographics. arXiv preprint arXiv:2209.14338.
  30. David Noever and Sam Hyams. 2023. Ai text-to-behavior: A study in steerability. arXiv preprint arXiv:2308.07326.
  31. OpenAI. 2022. Chatgpt - https://openai.com/blog/chatgpt#OpenAI.
  32. OpenAI. 2023. Gpt-4 technical report - https://cdn.openai.com/papers/gpt-4.pdf.
  33. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
  34. Keyu Pan and Yawen Zeng. 2023. Do llms possess a personality? making the mbti test an amazing evaluation for large language models. arXiv preprint arXiv:2307.16180.
  35. Improving language understanding by generative pre-training.
  36. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  37. Personality traits in large language models. arXiv preprint arXiv:2307.00184.
  38. Quantifying language models’ sensitivity to spurious features in prompt design or: How i learned to start worrying about prompt formatting. arXiv preprint arXiv:2310.11324.
  39. Have large language models developed a personality?: Applicability of self-assessment tests in measuring personality in llms. arXiv preprint arXiv:2305.14693.
  40. An attention-based denoising framework for personality detection in social media texts. arXiv preprint arXiv:2311.09945.
  41. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
  42. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  43. Orders are unwanted: Dynamic deep graph convolutional network for personality detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 13896–13904.
  44. Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068.
Citations (6)

Summary

  • The paper introduces a novel external evaluation method to assess LLM personality traits by predicting MBTI types from generated text responses.
  • It demonstrates that LLMs exhibit inconsistent personality profiles across roles, with notable differences between posts and comments.
  • The methodology leverages a fine-tuned Llama2-7B model achieving over 80% MBTI classification accuracy, outperforming traditional models.

LLM Personality Analysis via External Evaluation

The paper "Identifying Multiple Personalities in LLMs with External Evaluation" (2402.14805) introduces a novel methodology for assessing the personality traits of LLMs. This approach diverges from traditional self-assessment methods, which have been shown to be unreliable due to the sensitivity of LLMs to prompt variations and question ordering. Instead, the authors propose an external evaluation method, leveraging a fine-tuned Llama2-7B model to predict the MBTI personality type based on the LLM's generated text in response to open-ended situational questions. The paper's findings reveal a significant inconsistency in the personality traits exhibited by LLMs when generating posts versus comments, challenging the notion that LLMs possess a stable, enduring personality akin to that of humans.

Methodology

The methodology comprises three key stages, as illustrated in (Figure 1): Figure 1

Figure 1: This flowchart summarizes the steps in the methodology, from fine-tuning the Llama2-7B model to evaluating LLM personalities.

First, a Llama2-7B model is fine-tuned on a publicly available MBTI dataset to create a personality detection model. This fine-tuned model significantly outperforms existing state-of-the-art models in MBTI prediction accuracy, achieving over 80% accuracy in 16-class classification. The fine-tuning process employs LoRA, targeting the query and value layers. Second, the paper prompts LLMs (ChatGPT and Llama2 variants) to generate Twitter posts and comments based on summarized news events and existing tweets, respectively. This stage aims to elicit responses from LLMs in different roles, simulating real-world social media interactions. The prompts are designed to be simple to avoid influencing the LLMs' behavior. Finally, the generated text is fed into the fine-tuned personality detection model to obtain MBTI personality predictions. To mitigate sampling bias, 100 sets of 50 generations are created for each LLM and role, and the distribution of predicted MBTI types is analyzed.

Key Findings and Analysis

The paper's most striking finding is the significant difference in personality distributions exhibited by LLMs when generating posts versus comments. For example, the Llama2-7B model is predicted to have an ESTJ personality type when generating posts but an INFP type when generating comments. These inconsistencies are further visualized in (Figure 2), which shows the MBTI distribution of ChatGPT and Llama2 models. Figure 2

Figure 2: These pie charts show the MBTI distribution of ChatGPT and Llama2 models, using the external evaluation method, highlighting the personality differences between generating posts and comments.

To validate the external evaluation method, the authors conducted a similar analysis on human-generated posts and comments from eight celebrities. The results, shown in (Figure 3), demonstrate that the personality distributions for humans remain relatively consistent across the two roles, aligning with the psychological definition of personality as an enduring characteristic. Figure 3

Figure 3: These pie charts illustrate the MBTI distribution of four celebrities, using the external evaluation method, showing consistency between personality assessments from posts and comments.

This validation experiment confirms that the observed inconsistencies in LLM personalities are not artifacts of the personality detection model but rather reflect a fundamental difference in how LLMs and humans exhibit personality traits.

Implications and Future Directions

The findings of this paper have significant implications for how LLM behavior is understood and evaluated. The authors challenge the direct application of human personality frameworks to LLMs, arguing that the definition of personality may need to be re-evaluated in the context of artificial intelligence. The paper highlights the need for caution when using LLMs in applications where consistent and predictable behavior is critical. Future research should focus on developing new methods for defining and measuring personality in LLMs, taking into account their unique characteristics and capabilities. Additionally, further investigation into the factors that influence LLM personality shifts, such as prompt engineering and training data, could provide valuable insights into controlling and shaping LLM behavior.

Conclusion

The paper "Identifying Multiple Personalities in LLMs with External Evaluation" (2402.14805) presents a compelling case for re-evaluating the concept of personality in LLMs. By employing an external evaluation method and comparing LLM behavior to that of humans, the authors demonstrate that LLMs exhibit inconsistent personality traits across different contexts, challenging the notion that they possess a stable, enduring personality. This work calls for a more nuanced understanding of LLM behavior and the development of new frameworks for evaluating their personality traits.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.