Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 156 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 27 tok/s Pro
GPT-4o 110 tok/s Pro
Kimi K2 212 tok/s Pro
GPT OSS 120B 436 tok/s Pro
Claude Sonnet 4.5 39 tok/s Pro
2000 character limit reached

PersLLM: A Personified Training Approach for Large Language Models (2407.12393v5)

Published 17 Jul 2024 in cs.CL, cs.AI, and cs.CY

Abstract: LLMs exhibit human-like intelligence, enabling them to simulate human behavior and support various applications that require both humanized communication and extensive knowledge reserves. Efforts are made to personify LLMs with special training data or hand-crafted prompts, while correspondingly faced with challenges such as insufficient data usage or rigid behavior patterns. Consequently, personified LLMs fail to capture personified knowledge or express persistent opinion. To fully unlock the potential of LLM personification, we propose PersLLM, a framework for better data construction and model tuning. For insufficient data usage, we incorporate strategies such as Chain-of-Thought prompting and anti-induction, improving the quality of data construction and capturing the personality experiences, knowledge, and thoughts more comprehensively. For rigid behavior patterns, we design the tuning process and introduce automated DPO to enhance the specificity and dynamism of the models' personalities, which leads to a more natural opinion communication. Both automated metrics and expert human evaluations demonstrate the effectiveness of our approach. Case studies in human-machine interactions and multi-agent systems further suggest potential application scenarios and future directions for LLM personification.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  3. Charles S Carver. Perspectives on personality (7th Edition). Pearson, 2011.
  4. Benchmarking large language models in retrieval-augmented generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 17754–17762, 2024.
  5. Xueyong Chen. Lotus Lantern Poetry and Dreams by Lin Huiyin. 2nd Edition. People’s Literature Publishing House, 2012.
  6. Simulating opinion dynamics with networks of llm-based agents. arXiv preprint arXiv:2311.09618, 2023.
  7. Efficient and effective text encoding for chinese llama and alpaca. arXiv preprint arXiv:2304.08177, 2023. URL https://arxiv.org/abs/2304.08177.
  8. Enhancing chat language models by scaling high-quality instructional conversations. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 3029–3051, 2023.
  9. Iason Gabriel. Artificial intelligence, values, and alignment. Minds and machines, 30(3):411–437, 2020.
  10. Ai and the transformation of social science research. Science, 380(6650):1108–1109, 2023.
  11. S Güver and Renate Motschnig. Effects of diversity in teams and workgroups: A qulitative systematic review. International Journal of Business, Humanities and Technology, 2017.
  12. Pre-trained models: Past, present and future. AI Open, 2:225–250, 2021.
  13. Reference-free monolithic preference optimization with odds ratio. arXiv preprint arXiv:2403.07691, 2024.
  14. Minicpm: Unveiling the potential of small language models with scalable training strategies. arXiv preprint arXiv:2404.06395, 2024.
  15. Personallm: Investigating the ability of large language models to express personality traits. In Findings of the Association for Computational Linguistics: NAACL 2024, pages 3605–3627, 2024.
  16. The benefits, risks and bounds of personalizing the alignment of large language models to individuals. Nature Machine Intelligence, pages 1–10, 2024.
  17. Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
  18. Trustworthy llms: a survey and guideline for evaluating large language models’ alignment. arXiv preprint arXiv:2308.05374, 2023.
  19. Gardner Murphy. Personality: A biosocial approach to origins and structure. Harper & Brothers, 1947.
  20. Sylvia Nasar. A beautiful mind. Simon and Schuster, 2011.
  21. Large dual encoders are generalizable retrievers. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9844–9855, 2022.
  22. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744, 2022.
  23. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
  24. Influencing human–ai interaction by priming beliefs about ai can increase perceived trustworthiness, empathy and effectiveness. Nature Machine Intelligence, 5(10):1076–1086, 2023.
  25. What makes an ai device human-like? the role of interaction quality, empathy and perceived psychological anthropomorphic characteristics in the acceptance of artificial intelligence in the service industry. Computers in Human Behavior, 122:106855, 2021.
  26. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  27. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024.
  28. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
  29. Role play with large language models. Nature, 623(7987):493–498, 2023.
  30. Character-llm: A trainable agent for role-playing. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 13153–13187, 2023.
  31. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  32. A survey on large language model based autonomous agents. Frontiers of Computer Science, 18(6):1–26, 2024.
  33. Self-instruct: Aligning language models with self-generated instructions. In The 61st Annual Meeting Of The Association For Computational Linguistics, 2023a.
  34. Aligning large language models with human: A survey. arXiv preprint arXiv:2307.12966, 2023b.
  35. Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. arXiv preprint arXiv:2310.00746, 2023c.
  36. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022.
  37. Simple synthetic data reduces sycophancy in large language models. arXiv e-prints, pages arXiv–2308, 2023.
  38. Diyi Yang. Human-ai interaction in the age of large language models. In Proceedings of the AAAI Symposium Series, volume 3, pages 66–67, 2024.
  39. Exploring the impact of instruction data scaling on large language models: An empirical study on real-world use cases. arXiv preprint arXiv:2303.14742, 2023.
  40. Knowledgeable preference alignment for llms in domain-specific question answering. arXiv preprint arXiv:2311.06503, 2023.
  41. A survey of large language models. arXiv preprint arXiv:2303.18223, 2023.
  42. Characterglm: Customizing chinese conversational ai characters with large language models. arXiv preprint arXiv:2311.16832, 2023a.
  43. Sotopia: Interactive evaluation for social intelligence in language agents. In The 12th International Conference on Learning Representations, 2023b.
Citations (2)

Summary

  • The paper presents PersLLM, which integrates personality traits into LLM training to create more authentic and dynamic human-like interactions.
  • It employs psychology-grounded data construction and Direct Preference Optimization to embed consistent and evolving personality characteristics.
  • Experiments in single-agent, multi-agent, and human-agent settings demonstrate improved interaction quality and realistic simulation of distinct personalities.

PersLLM: A Personified Training Approach for LLMs

The paper "PersLLM: A Personified Training Approach for LLMs" explores the integration of personality traits into LLMs to enhance their utility in various applications, such as social simulations, human-machine interactions, and collaborative systems. The authors recognize the necessity for LLMs to exhibit distinct personality characteristics to improve their practical applicability, addressing the uniform response patterns often seen in current models.

Introduction

The authors identify a deficiency in current LLMs: a lack of personified traits that would allow them to simulate human interactions authentically. They propose a personified training methodology called PersLLM, which incorporates psychology-grounded principles such as social practice, consistency, and dynamic development directly into LLM parameters, thereby enhancing personality traits. Figure 1

Figure 1: Schematic diagram for PersLLM's personified training approach, displaying data collection, annotation, and tuning strategies.

Methodology

Personified Data Construction

The construction of personified data involves collecting raw data on target personalities. This includes biographies, personal writings, and third-party descriptions that offer insights into an individual's experiences, knowledge, and speech style. The raw data is then annotated through conversational tuning strategies, including the use of temporal labels for dynamic development, anti-induced data to maintain consistency, and Chain-of-Thought (CoT) prompting to ensure depth in reasoning.

Model Training

The annotated conversational data is integrated into model training, where personality features are embedded into the LLM parameters rather than relying solely on external prompt engineering or retrieval-augmented generation (RAG). The approach uses automatic Direct Preference Optimization (DPO) to differentiate responses across different temporal stages and personalities.

Experiments and Results

Single-agent Evaluation

The paper includes a single-agent evaluation using the Harry Potter dataset, demonstrating that PersLLM generates responses more aligned with distinct personalities compared to baseline methods. The authors utilize BLEU and ROUGE metrics to assess the model's performance in text generation tasks. Figure 2

Figure 2: Comparison of attitudes, knowledge reserves, and temporal changes in personalities of personified models versus role-play instruction tuning models.

Multi-agent Communication

PersLLM shows efficacy in simulating human-like interactions in multi-agent settings, particularly in maintaining opinion consistency and fostering collaborative creativity among agents. Conflict and cooperation instances are tested among different personality models, such as Harry and Voldemort, and John Nash and Huiyin Lin. Figure 3

Figure 3: Conflict between Harry Potter and Lord Voldemort showcasing PersLLM's ability to maintain distinct personalities and prevent opinion convergence.

Figure 4

Figure 4: Cooperation between John Nash and Huiyin Lin, illustrating PersLLM's capability to enable specialized interdisciplinary dialogues.

Human-Agent Interaction

Human-agent interaction tests suggest that personified models provide a superior interactive experience compared to generic models. The evaluations, conducted with real participants, highlight improvements in trust, companionship, character simulation, and user engagement. Figure 5

Figure 5: Evaluation of human interaction with the PersLLM Huiyin Lin agent, showing enhanced satisfaction in multiple metrics.

Conclusion

The PersLLM framework successfully models dynamic personality traits in LLMs, showing potential benefits for applications in human-like conversational agents and simulated social research. The personified approach emphasizes consistency, adaptability, and the realistic portrayal of complex human interactions across various contexts. Future work will aim to improve real-time personality modeling and address ethical concerns related to privacy and the misuse of personification technologies.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 3 tweets and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com