Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 43 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 17 tok/s Pro
GPT-5 High 19 tok/s Pro
GPT-4o 96 tok/s Pro
Kimi K2 197 tok/s Pro
GPT OSS 120B 455 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

NLP at UC Santa Cruz at SemEval-2024 Task 5: Legal Answer Validation using Few-Shot Multi-Choice QA (2404.03150v1)

Published 4 Apr 2024 in cs.CL and cs.AI

Abstract: This paper presents our submission to the SemEval 2024 Task 5: The Legal Argument Reasoning Task in Civil Procedure. We present two approaches to solving the task of legal answer validation, given an introduction to the case, a question and an answer candidate. Firstly, we fine-tuned pre-trained BERT-based models and found that models trained on domain knowledge perform better. Secondly, we performed few-shot prompting on GPT models and found that reformulating the answer validation task to be a multiple-choice QA task remarkably improves the performance of the model. Our best submission is a BERT-based model that achieved the 7th place out of 20.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (14)
  1. Longformer: The Long-Document Transformer.
  2. The Legal Argument Reasoning Task in Civil Procedure.
  3. Language Models are Few-Shot Learners.
  4. Principled Instructions Are All You Need for Questioning LLaMA-1/2, GPT-3.5/4.
  5. LEGAL-BERT: The Muppets straight out of Law School. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 2898–2904, Online. Association for Computational Linguistics.
  6. LeXFfiles and LegalLAMA: Facilitating English Multinational Legal Language Model Development. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15513–15535, Toronto, Canada. Association for Computational Linguistics.
  7. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
  8. Glannon, J. W. (2019). The Glannon Guide to Civil Procedure. Wolters Kluwer, New York, NY, 4 edition.
  9. A Free Format Legal Question Answering System. In Proceedings of the Natural Legal Language Processing Workshop 2021, pages 107–113, Punta Cana, Dominican Republic. Association for Computational Linguistics.
  10. OpenAI (2023a). GPT-3.5 Model Documentation. https://platform.openai.com/docs/models/gpt-3-5-turbo. Accessed: 2024-02-05.
  11. OpenAI (2023b). GPT-4 Model Documentation. https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo. Accessed: 2024-02-05.
  12. Boosting methods for multi-class imbalanced data classification: an experimental review. Journal of Big Data, 7:1–47.
  13. A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT.
  14. When Does Pretraining Help? Assessing Self-Supervised Learning for Law and the CaseHOLD Dataset.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.