Zero-Shot and Few-Shot Learning for Dynamic Knowledge Base Updates in Chatbots

Main Article Content

Srikanth Balla
Dr Anand Singh

Abstract

Chatbots are increasingly dependent on dynamic knowledge bases to generate on-topic, context-aware responses in real-time conversations. Historically, these knowledge bases must be updated from large annotated datasets and constant retraining by hand, which poses scalability and responsiveness issues in fast-changing domains. Recent breakthroughs in zero-shot and few-shot learning have been promising to enable models to generalize to new tasks or classes with limited or no additional labeled data. It remains an underresearched area, however, to apply such approaches to dynamic updates in chatbot knowledge bases. Current studies are primarily on static knowledge bases or require extensive fine-tuning, and thus they do not fit to solve the problem related to continual updates and domain change. This research fills a critical gap in integrating zero-shot and few-shot learning methods to allow chatbots to update their knowledge bases autonomously and efficiently with little supervision. This research introduces a new framework that leverages zero-shot and few-shot learning methods to dynamically learn, extract, and add new knowledge to chatbot systems. With the integration of transfer learning and context awareness, the framework seeks to alleviate reliance on large labeled datasets and manual tuning. Our method encourages more adaptive, scalable, and resilient chatbot behavior in changing environments. We test the framework on multiple real-world conversational datasets and show substantial improvements in update speed, accuracy, and generalization over traditional approaches. This research makes a contribution to the field of chatbot knowledge management by bridging the gap between new learning paradigms and requirements for real-world deployment, ultimately resulting in more intelligent and self-sustaining conversational agents.

Article Details

How to Cite
Balla, S., & Singh, D. A. (2024). Zero-Shot and Few-Shot Learning for Dynamic Knowledge Base Updates in Chatbots. Journal of Quantum Science and Technology (JQST), 1(1), Feb(104–118). Retrieved from https://jqst.org/index.php/j/article/view/298
Section
Original Research Articles

References

• Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., ... & Amodei, D. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877–1901.

• Chen, Y., Chang, J., & Zettlemoyer, L. (2020). Meta-learning for low-resource natural language generation. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2324–2334. https://doi.org/10.18653/v1/2020.emnlp-main.185

• Sandeep Dommari. (2023). The Intersection of Artificial Intelligence and Cybersecurity: Advancements in Threat Detection and Response. International Journal for Research Publication and Seminar, 14(5), 530–545. https://doi.org/10.36676/jrps.v14.i5.1639

• Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT, 4171–4186. https://doi.org/10.18653/v1/N19-1423

• Gao, T., Fisch, A., & Chen, D. (2021). Making pre-trained language models better few-shot learners. ACL 2021, 3816–3830. https://doi.org/10.18653/v1/2021.acl-long.296

• Howard, J., & Ruder, S. (2018). Universal language model fine-tuning for text classification. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 328–339. https://doi.org/10.18653/v1/P18-1031

• Kirkpatrick, J., Pascanu, R., Rabinowitz, N., Veness, J., Desjardins, G., Rusu, A. A., ... & Hadsell, R. (2017). Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences, 114(13), 3521–3526. https://doi.org/10.1073/pnas.1611835114

• Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., ... & Riedel, S. (2020). Retrieval-augmented generation for knowledge-intensive NLP tasks. Advances in Neural Information Processing Systems, 33, 9459–9474.

• Lin, J., Li, M., & Zhang, X. (2022). Zero-shot domain adaptation for dialogue systems using pretrained language models. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 8474–8481. https://doi.org/10.1609/aaai.v36i8.21094

• Patel, S., Singh, R., & Kumar, A. (2023). Benchmarking zero-shot and few-shot learning for dynamic knowledge updates in conversational AI. Journal of Artificial Intelligence Research, 78, 1023–1045. https://doi.org/10.1613/jair.1.13257

• Serban, I. V., Lowe, R., Charlin, L., & Pineau, J. (2016). Generative deep neural networks for dialogue: A short review. Proceedings of the NIPS 2016 Workshop on Machine Learning for Spoken Language Understanding and Interaction.

• Shum, H.-Y., He, X., & Li, D. (2018). From Eliza to XiaoIce: Challenges and opportunities with social chatbots. Frontiers of Information Technology & Electronic Engineering, 19(1), 10–26. https://doi.org/10.1631/FITEE.1700826

• Sun, Y., Chen, Z., & Liu, Y. (2022). Continual learning for conversational agents: A review and perspective. IEEE Transactions on Neural Networks and Learning Systems, 33(7), 2951–2966. https://doi.org/10.1109/TNNLS.2021.3059143

• Wang, X., Yu, D., & Liu, Y. (2018). Zero-shot learning: A comprehensive evaluation of the good, the bad and the ugly. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(12), 2987–3002. https://doi.org/10.1109/TPAMI.2018.2836436

• Dommari, S., & Khan, S. (2023). Implementing Zero Trust Architecture in cloud-native environments: Challenges and best practices. International Journal of All Research Education and Scientific Methods (IJARESM), 11(8), 2188. Retrieved from http://www.ijaresm.com

• Wang, Y., Li, F., & Zhao, J. (2024). Automated knowledge verification for conversational AI: A trustworthiness framework. Information Processing & Management, 61(2), 102987. https://doi.org/10.1016/j.ipm.2023.102987

• Xie, T., Chen, Z., & Sun, M. (2021). Few-shot learning with knowledge graph embeddings for conversational AI. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 6353–6364. https://doi.org/10.18653/v1/2021.emnlp-main.504

• Zhang, L., Wu, Y., & Wang, S. (2023). Dynamic knowledge base updating for dialogue systems with few-shot and continual learning. Knowledge-Based Systems, 263, 109682. https://doi.org/10.1016/j.knosys.2023.109682

• Zhou, Q., Yang, F., & Chen, H. (2023). Multimodal few-shot learning for dynamic knowledge integration in chatbots. IEEE Transactions on Multimedia, 25(4), 1024–1035. https://doi.org/10.1109/TMM.2023.3245127