Chatlaw: A Multi-Agent Collaborative Legal Assistant with Knowledge Graph Enhanced Mixture-of-Experts Large Language Model (2306.16092v2)
Abstract: AI legal assistants based on LLMs can provide accessible legal consulting services, but the hallucination problem poses potential legal risks. This paper presents Chatlaw, an innovative legal assistant utilizing a Mixture-of-Experts (MoE) model and a multi-agent system to enhance the reliability and accuracy of AI-driven legal services. By integrating knowledge graphs with artificial screening, we construct a high-quality legal dataset to train the MoE model. This model utilizes different experts to address various legal issues, optimizing the accuracy of legal responses. Additionally, Standardized Operating Procedures (SOP), modeled after real law firm workflows, significantly reduce errors and hallucinations in legal services. Our MoE model outperforms GPT-4 in the Lawbench and Unified Qualification Exam for Legal Professionals by 7.73% in accuracy and 11 points, respectively, and also surpasses other models in multiple dimensions during real-case consultations, demonstrating our robust capability for legal consultation.
- Falcon-40B: an open large language model with state-of-the-art performance. 2023.
- Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality, March 2023.
- LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
- Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547, 2019.
- OpenAI. Gpt-4 technical report, 2023.
- Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face, 2023.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
- Huatuo: Tuning llama model with chinese medical knowledge, 2023.
- Bloomberggpt: A large language model for finance, 2023.
- Fingpt: Open-source financial large language models, 2023.
- Zero-shot learners for natural language understanding via a unified multiple choice perspective, 2022.
- GLM-130b: An open bilingual pre-trained model. In The Eleventh International Conference on Learning Representations (ICLR), 2023.
- Judging llm-as-a-judge with mt-bench and chatbot arena, 2023.
- Chatmed: A chinese medical large language model. https://github.com/michael-wzhu/ChatMed, 2023.