Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Injecting Domain Knowledge in Language Models for Task-Oriented Dialogue Systems (2212.08120v1)

Published 15 Dec 2022 in cs.CL and cs.AI

Abstract: Pre-trained LLMs (PLM) have advanced the state-of-the-art across NLP applications, but lack domain-specific knowledge that does not naturally occur in pre-training data. Previous studies augmented PLMs with symbolic knowledge for different downstream NLP tasks. However, knowledge bases (KBs) utilized in these studies are usually large-scale and static, in contrast to small, domain-specific, and modifiable knowledge bases that are prominent in real-world task-oriented dialogue (TOD) systems. In this paper, we showcase the advantages of injecting domain-specific knowledge prior to fine-tuning on TOD tasks. To this end, we utilize light-weight adapters that can be easily integrated with PLMs and serve as a repository for facts learned from different KBs. To measure the efficacy of proposed knowledge injection methods, we introduce Knowledge Probing using Response Selection (KPRS) -- a probe designed specifically for TOD models. Experiments on KPRS and the response generation task show improvements of knowledge injection with adapters over strong baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Denis Emelin (5 papers)
  2. Daniele Bonadiman (10 papers)
  3. Sawsan Alqahtani (7 papers)
  4. Yi Zhang (994 papers)
  5. Saab Mansour (32 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.