VerifiNER: Verification-augmented NER via Knowledge-grounded Reasoning with Large Language Models (2402.18374v2)
Abstract: Recent approaches in domain-specific named entity recognition (NER), such as biomedical NER, have shown remarkable advances. However, they still lack of faithfulness, producing erroneous predictions. We assume that knowledge of entities can be useful in verifying the correctness of the predictions. Despite the usefulness of knowledge, resolving such errors with knowledge is nontrivial, since the knowledge itself does not directly indicate the ground-truth label. To this end, we propose VerifiNER, a post-hoc verification framework that identifies errors from existing NER methods using knowledge and revises them into more faithful predictions. Our framework leverages the reasoning abilities of LLMs to adequately ground on knowledge and the contextual information in the verification process. We validate effectiveness of VerifiNER through extensive experiments on biomedical datasets. The results suggest that VerifiNER can successfully verify errors from existing models as a model-agnostic approach. Further analyses on out-of-domain and low-resource settings show the usefulness of VerifiNER on real-world applications.
- Dhananjay Ashok and Zachary C Lipton. 2023. Promptner: Prompting for named entity recognition. arXiv preprint arXiv:2305.15444.
- O. Bodenreider. 2004. The unified medical language system (umls): integrating biomedical terminology. Nucleic Acids Research, 32(90001):267D – 270.
- Improving language models by retrieving from trillions of tokens. In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pages 2206–2240. PMLR.
- Dialogue chain-of-thought distillation for commonsense-aware conversational agents. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 5606–5632, Singapore. Association for Computational Linguistics.
- Data augmentation for cross-domain named entity recognition. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 5346–5356, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Xiang Dai. 2021. Recognising biomedical names: Challenges and solutions. arXiv preprint arXiv:2106.12230.
- Biomedical named entity recognition via dictionary-based synonym generalization. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 14621–14635, Singapore. Association for Computational Linguistics.
- Retrieval augmented language model pre-training. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 3929–3938. PMLR.
- Few-shot learning with retrieval augmented language models. CoRR, abs/2208.03299.
- Minbyul Jeong and Jaewoo Kang. 2023. Consistency enhancement of model prediction on document-level named entity recognition. Bioinformatics, 39(6).
- KALA: knowledge-augmented language model adaptation. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July 10-15, 2022, pages 5144–5167. Association for Computational Linguistics.
- Explainable ai for bioinformatics: Methods, tools and applications. Briefings in bioinformatics, 24(5):bbad236.
- Genia corpus—a semantically annotated corpus for bio-textmining. Bioinformatics, 19(suppl1):i180–i182.
- Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems, volume 35, pages 22199–22213. Curran Associates, Inc.
- Neural architectures for named entity recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 260–270.
- Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234–1240.
- Retrieval-augmented generation for knowledge-intensive NLP tasks. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual.
- Large language models with controllable working memory. arXiv preprint arXiv:2211.05110.
- Biocreative v cdr task corpus: a resource for chemical disease relation extraction. Database, 2016:baw068.
- Towards improving neural named entity recognition with gazetteers. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5301–5307, Florence, Italy. Association for Computational Linguistics.
- Coarse-to-Fine Pre-training for Named Entity Recognition. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6345–6354, Online. Association for Computational Linguistics.
- Named entity recognition without gazetteers. In Ninth Conference of the European Chapter of the Association for Computational Linguistics, pages 1–8.
- Knowledge-aware named entity recognition with alleviating heterogeneity. Proceedings of the AAAI Conference on Artificial Intelligence, 35(15):13595–13603.
- OpenAI. 2023. Chatgpt. https://openai.com/blog/chatgpt.
- ERICA: improving entity and relation understanding for pre-trained language models via contrastive learning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, pages 3350–3363. Association for Computational Linguistics.
- Gpt-ner: Named entity recognition via large language models. arXiv preprint arXiv:2304.10428.
- Improving named entity recognition by external context retrieving and cooperative learning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1800–1812, Online. Association for Computational Linguistics.
- Self-consistency improves chain of thought reasoning in language models. In The Eleventh International Conference on Learning Representations.
- Improving named entity recognition via bridge-based domain adaptation. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3869–3882, Toronto, Canada. Association for Computational Linguistics.
- LUKE: deep contextualized entity representations with entity-aware self-attention. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pages 6442–6454. Association for Computational Linguistics.
- FactMix: Using a few labeled in-domain examples to generalize to cross-domain named entity recognition. In Proceedings of the 29th International Conference on Computational Linguistics, pages 5360–5371, Gyeongju, Republic of Korea. International Committee on Computational Linguistics.
- PDALN: Progressive domain adaptation over a pre-trained model for low-resource cross-domain named entity recognition. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 5441–5451, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Siren’s song in the ai ocean: a survey on hallucination in large language models. arXiv preprint arXiv:2309.01219.
- ERNIE: enhanced language representation with informative entities. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers, pages 1441–1451. Association for Computational Linguistics.
- MELM: Data augmentation with masked entity language modeling for low-resource NER. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2251–2262, Dublin, Ireland. Association for Computational Linguistics.
- Seoyeon Kim (7 papers)
- Kwangwook Seo (4 papers)
- Hyungjoo Chae (18 papers)
- Jinyoung Yeo (46 papers)
- Dongha Lee (63 papers)