KRGP: knowledge-based response generation with persona
Zapiski Nauchnykh Seminarov POMI, Investigations on applied mathematics and informatics. Part II–1, Tome 529 (2023), pp. 72-85
Cet article a éte moissonné depuis la source Math-Net.Ru

Voir la notice du chapitre de livre

To create a personalized response, a generative model must take into account personal information about the user, question asked, and domain knowledge. Therefore, it is necessary to learn how to extract relevant information that will help the generative model to compose a response to the user. In this work, we propose to split the process into three stages: selection of relevant sentences from the textual knowledge base, selection of the most suitable sentences of the textual persona description taking into account the extracted knowledge, and response generation based on the knowledge and persona. We use the sentence Transformer and adapt the algorithm from the CLIP paper to obtain contextual sentence embeddings to extract the most relevant text spans from the knowledge base. We have found that the focal loss shows better results in tasks of binary classification of a persona using the FoCus imbalanced dataset as an example. We have also shown that text2text Transformer BART performs well in the tasks of conditional response generation in a dialog. This system achieved a state-of-the-art result at the leaderboard of The 1st Workshop on Customized Chat Grounding Personahttps://sites.google.com/view/persona-knowledge-workshop/home. The code for this work is available at https://github.com/dmitrymailk/deeppavlov_focus.
@article{ZNSL_2023_529_a5,
     author = {D. Kosenko and D. Zharikova},
     title = {KRGP: knowledge-based response generation with persona},
     journal = {Zapiski Nauchnykh Seminarov POMI},
     pages = {72--85},
     year = {2023},
     volume = {529},
     language = {en},
     url = {http://geodesic.mathdoc.fr/item/ZNSL_2023_529_a5/}
}
TY  - JOUR
AU  - D. Kosenko
AU  - D. Zharikova
TI  - KRGP: knowledge-based response generation with persona
JO  - Zapiski Nauchnykh Seminarov POMI
PY  - 2023
SP  - 72
EP  - 85
VL  - 529
UR  - http://geodesic.mathdoc.fr/item/ZNSL_2023_529_a5/
LA  - en
ID  - ZNSL_2023_529_a5
ER  - 
%0 Journal Article
%A D. Kosenko
%A D. Zharikova
%T KRGP: knowledge-based response generation with persona
%J Zapiski Nauchnykh Seminarov POMI
%D 2023
%P 72-85
%V 529
%U http://geodesic.mathdoc.fr/item/ZNSL_2023_529_a5/
%G en
%F ZNSL_2023_529_a5
D. Kosenko; D. Zharikova. KRGP: knowledge-based response generation with persona. Zapiski Nauchnykh Seminarov POMI, Investigations on applied mathematics and informatics. Part II–1, Tome 529 (2023), pp. 72-85. http://geodesic.mathdoc.fr/item/ZNSL_2023_529_a5/

[1] M. Adam, M. Wessel, A. Benlian, “Ai-based chatbots in customer service and their effects on user compliance”, Electronic Markets, 31:2 (2021), 427–445

[2] E. Dinan, S. Roller, K. Shuster, A. Fan, M. Auli, J. Weston, Wizard of wikipedia: Knowledge-powered conversational agents, 2018, arXiv: 1811.01241

[3] C. Gao, W. Lei, X. He, M. de Rijke, Tat-Seng Chua, “Advances and challenges in conversational recommender systems: A survey”, AI Open, 2 (2021), 100–126

[4] P. He, J. Gao, W. Chen, Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing, 2021

[5] Y. Jang, J. Lim, Y. Hur, D. Oh, S. Son, Y. Lee, D. Shin, S. Kim, H. Lim, “Call for customized conversation: Customized conversation grounding persona and knowledge”, Proceedings of the AAAI Conference on Artificial Intelligence, 36 (2022), 10803–10812

[6] H. Jiang, Y. Cheng, J. Yang, S. Gao, “Ai-powered chatbot communication with customers: Dialogic interactions, satisfaction, engagement, and customer behavior”, Computers in Human Behavior, 134 (2022), 107329

[7] V. Konovalov, O. Melamud, R. Artstein, I. Dagan, “Collecting Better Training Data using Biased Agent Policies in Negotiation Dialogues”, Proceedings of WOCHAT, the Second Workshop on Chatbots and Conversational Agent Technologies (Los Angeles, Zerotype, September 2016)

[8] M. Lewis, Y. Liu, N. Goyal, M. Ghazvininejad, A. Mohamed, O. Levy, V. Stoyanov, L. Zettlemoyer, Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, 2019, arXiv: 1910.13461

[9] Y. Li, S. A. Hayati, W. Shi, Z. Yu, Deux: an attribute-guided framework for sociable recommendation dialog systems, 2021, arXiv: 2105.00825

[10] J. Lim, M. Kang, Y. Hur, S. Jung, J. Kim, Y. Jang, D. Lee, H. Ji, D. Shin, S. Kim, et al., You truly understand what i need: Intellectual and friendly dialogue agents grounding knowledge and persona, 2023, arXiv: 2301.02401

[11] C.-Y. Lin, “ROUGE: A package for automatic evaluation of summaries”, Text Summarization Branches Out (Barcelona, Spain, July 2004), Association for Computational Linguistics, 74–81

[12] T.-Yi Lin, P. Goyal, R. Girshick, K. He, P. Dollár, “Focal loss for dense object detection”, Proceedings of the IEEE international conference on computer vision, 2017, 2980–2988

[13] I. Loshchilov, F. Hutter, Decoupled weight decay regularization, 2017, arXiv: 1711.05101

[14] K. K. Pal, K. Kashihara, U. Anantheswaran, K. C. Kuznia, S. Jagtap, C. Baral, Exploring the limits of transfer learning with unified model in the cybersecurity domain, 2023, arXiv: 2302.10346

[15] K. Papineni, S. Roukos, T. Ward, W. Jing Zhu, “Bleu: a method for automatic evaluation of machine translation”, Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 2002, 311–318

[16] M. Popović, “chrF: character n-gram F-score for automatic MT evaluation”, Proceedings of the Tenth Workshop on Statistical Machine Translation (Lisbon, Portugal), Association for Computational Linguistics, 392–395

[17] A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al., “Learning transferable visual models from natural language supervision”, International conference on machine learning, PMLR 2021, 8748–8763

[18] N. Reimers, I. Gurevych, “Sentence-bert: Sentence embeddings using siamese bert-networks”, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (11 2019), Association for Computational Linguistics

[19] S. Roller, E. Dinan, N. Goyal, D. Ju, M. Williamson, Y. Liu, J. Xu, M. Ott, K. Shuster, E. M. Smith, et al., Recipes for building an open-domain chatbot, 2020, arXiv: 2004.13637

[20] S. Saha, S. Das, R. K. Srihari, “Proto-gen: An end-to-end neural generator for persona and knowledge grounded response generation”, Proceedings of the 1st Workshop on Customized Chat Grounding Persona and Knowledge, 2022, 9–14

[21] K. Shuster, S. Poff, M. Chen, D. Kiela, J. Weston, Retrieval augmentation reduces hallucination in conversation, 2021, arXiv: 2104.07567

[22] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf, M. Funtowicz, et al., Huggingface's transformers: State-of-the-art natural language processing, 2019, arXiv: 1910.03771

[23] C. Xu, P. Li, W. Wang, H. Yang, S. Wang, C. Xiao, “Cosplay: Concept set guided personalized dialogue generation across both party personas”, Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2022, 201–211

[24] S. Zhang, E. Dinan, J. Urbanek, A. Szlam, D. Kiela, J. Weston, Personalizing dialogue agents: I have a dog, do you have pets too?, 2018, arXiv: 1801.07243

[25] H. Zhong, Z. Dou, Y. Zhu, H. Qian, J.-R. Wen, Less is more: Learning to refine dialogue history for personalized dialogue generation, 2022, arXiv: 2204.08128

[26] K. Zhou, S. Prabhumoye, A. W. Black, A dataset for document grounded conversations, 2018, arXiv: 1809.07358