Research Area:  Machine Learning
As an intelligent way to interact with computers, the dialog system has been catching more and more attention. However, most research efforts only focus on text-based dialog systems, completely ignoring the rich semantics conveyed by the visual cues. Indeed, the desire for multimodal task-oriented dialog systems is growing with the rapid expansion of many domains, such as the online retailing and travel. Besides, few work considers the hierarchical product taxonomy and the users attention to products explicitly. The fact is that users tend to express their attention to the semantic attributes of products such as color and style as the dialog goes on. Towards this end, in this work, we present a hierarchical User attention-guided Multimodal Dialog system, named UMD for short. UMD leverages a bidirectional Recurrent Neural Network to model the ongoing dialog between users and chatbots at a high level; As to the low level, the multimodal encoder and decoder are capable of encoding multimodal utterances and generating multimodal responses, respectively. The multimodal encoder learns the visual presentation of images with the help of a taxonomy-attribute combined tree, and then the visual features interact with textual features through an attention mechanism; whereas the multimodal decoder selects the required visual images and generates textual responses according to the dialog history. To evaluate our proposed model, we conduct extensive experiments on a public multimodal dialog dataset in the retailing domain. Experimental results demonstrate that our model outperforms the existing state-of-the-art methods by integrating the multimodal utterances and encoding the visual features based on the users attribute-level attention.
Keywords:  
User Attention-guided
Multimodal
Dialog Systems
Deep Learning
Machine Learning
Author(s) Name:  Chen Cui , Wenjie Wang , Xuemeng Song , Minlie Huang , Xin-Shun Xu , Liqiang Nie
Journal name:  
Conferrence name:  SIGIR-19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval
Publisher name:  ACM
DOI:  10.1145/3331184.3331226
Volume Information:  Pages 445–454
Paper Link:   https://dl.acm.org/doi/abs/10.1145/3331184.3331226