Enhancing Knowledge Retrieval with Topic Modeling for Knowledge-Grounded Dialogue

Nhat Tran, Diane Litman


Abstract
Knowledge retrieval is one of the major challenges in building a knowledge-grounded dialogue system. A common method is to use a neural retriever with a distributed approximate nearest-neighbor database to quickly find the relevant knowledge sentences. In this work, we propose an approach that utilizes topic modeling on the knowledge base to further improve retrieval accuracy and as a result, improve response generation. Additionally, we experiment with a large language model (LLM), ChatGPT, to take advantage of the improved retrieval performance to further improve the generation results. Experimental results on two datasets show that our approach can increase retrieval and generation performance. The results also indicate that ChatGPT is a better response generator for knowledge-grounded dialogue when relevant knowledge is provided.
Anthology ID:
2024.lrec-main.530
Volume:
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Month:
May
Year:
2024
Address:
Torino, Italia
Editors:
Nicoletta Calzolari, Min-Yen Kan, Veronique Hoste, Alessandro Lenci, Sakriani Sakti, Nianwen Xue
Venues:
LREC | COLING
SIG:
Publisher:
ELRA and ICCL
Note:
Pages:
5986–5995
Language:
URL:
https://aclanthology.org/2024.lrec-main.530
DOI:
Bibkey:
Cite (ACL):
Nhat Tran and Diane Litman. 2024. Enhancing Knowledge Retrieval with Topic Modeling for Knowledge-Grounded Dialogue. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 5986–5995, Torino, Italia. ELRA and ICCL.
Cite (Informal):
Enhancing Knowledge Retrieval with Topic Modeling for Knowledge-Grounded Dialogue (Tran & Litman, LREC-COLING 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.lrec-main.530.pdf