Character-LLM: A Trainable Agent for Role-Playing (2310.10158v2)
Abstract: LLMs can be used to serve as agents to simulate human behaviors, given the powerful ability to understand human instructions and provide high-quality generated texts. Such ability stimulates us to wonder whether LLMs can simulate a person in a higher form than simple human behaviors. Therefore, we aim to train an agent with the profile, experience, and emotional states of a specific person instead of using limited prompts to instruct ChatGPT API. In this work, we introduce Character-LLM that teach LLMs to act as specific people such as Beethoven, Queen Cleopatra, Julius Caesar, etc. Our method focuses on editing profiles as experiences of a certain character and training models to be personal simulacra with these experiences. To assess the effectiveness of our approach, we build a test playground that interviews trained agents and evaluates whether the agents \textit{memorize} their characters and experiences. Experimental results show interesting observations that help build future simulacra of humankind.
- Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862.
- Joseph Bates. 1994. The role of emotion in believable agents. Commun. ACM, 37(7):122–125.
- On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258.
- The cog project: Building a humanoid robot. In Computation for Metaphors, Analogy, and Agents, number 1562 in Lecture Notes on Artificial Intelligence, pages 52–87, Berlin. Springer-Verlag.
- Language models are few-shot learners. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual.
- Dungeons and dragons as a dialog challenge for artificial intelligence. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9379–9393, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality.
- Jonas Freiknecht and Wolfgang Effelsberg. 2020. Procedural generation of interactive stories using language models. In International Conference on the Foundations of Digital Games (FDG ’20), page 8, Bugibba, Malta. ACM.
- A survey on automated fact-checking. Trans. Assoc. Comput. Linguistics, 10:178–206.
- Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1–38.
- Soda: Million-scale dialogue distillation with social commonsense contextualization. arXiv preprint arXiv:2212.10465.
- Evaluating the factual consistency of abstractive text summarization. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pages 9332–9346. Association for Computational Linguistics.
- John Laird and Michael VanLent. 2001. Human-level ai’s killer application: Interactive computer games. AI Magazine, 22(2):15.
- Neil Madden and Brian Logan. 2007. Collaborative narrative generation in persistent virtual environments. In Intelligent Narrative Technologies, Papers from the 2007 AAAI Fall Symposium, Arlington, Virginia, USA, November 9-11, 2007, volume FS-07-05 of AAAI Technical Report, pages 71–78. AAAI Press.
- Developing game ai agent behaving like human by mixing reinforcement learning and supervised learning. In Proceedings of the 18th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD), pages 153–158, Kanazawa, Japan.
- Deep learning for deepfakes creation and detection: A survey. Computer Vision and Image Understanding, 223:103525.
- OpenAI. 2023. GPT-4 technical report. CoRR, abs/2303.08774.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442.
- Social simulacra: Creating populated prototypes for social computing systems. In In the 35th Annual ACM Symposium on User Interface Software and Technology (UIST ’22), UIST ’22, New York, NY, USA. Association for Computing Machinery.
- Mark O. Riedl. 2012. Interactive narrative: A novel application of artificial intelligence for computer games. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence (AAAI’12), pages 2160–2165.
- Mark O. Riedl and R. Michael Young. 2005. An objective character believability evaluation procedure for multi-agent story generation systems. In Proceedings of the 5th International Working Conference on Intelligent Virtual Agents (IVA’05), pages 58–70, Kos, Greece.
- Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
- Frank Thomas and Ollie Johnston. 1981. Disney Animation: The Illusion of Life. Abbeville Press, New York.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
- Neural codec language models are zero-shot text to speech synthesizers. arXiv preprint arXiv:2301.02111.
- Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291.
- Self-instruct: Aligning language model with self generated instructions. arXiv preprint arXiv:2212.10560.
- Hififace: 3d shape and semantic prior guided high fidelity face swapping. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19-27 August 2021, pages 1136–1142. ijcai.org.
- Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
- Promptchainer: Chaining large language model prompts through visual programming. In CHI EA ’22: Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems.
- Ai chains: Transparent and controllable human-ai interaction by chaining large language model prompts. In CHI ’22: Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems.
- Wizardlm: Empowering large language models to follow complex instructions. arXiv preprint arXiv:2304.12244.
- Speak foreign languages with your own voice: Cross-lingual neural codec language modeling. arXiv preprint arXiv:2303.03926.
- Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206.