Amazing technological breakthrough possible @S-Logix pro@slogix.in

Office Address

  • #5, First Floor, 4th Street Dr. Subbarayan Nagar Kodambakkam, Chennai-600 024 Landmark : Samiyar Madam
  • pro@slogix.in
  • +91- 81240 01111

Social List

User Attention-guided Multimodal Dialog Systems - 2019

User attention-guided multimodal dialog systems

Research paper on User Attention-guided Multimodal Dialog Systems

Research Area:  Machine Learning

Abstract:

As an intelligent way to interact with computers, the dialog system has been catching more and more attention. However, most research efforts only focus on text-based dialog systems, completely ignoring the rich semantics conveyed by the visual cues. Indeed, the desire for multimodal task-oriented dialog systems is growing with the rapid expansion of many domains, such as the online retailing and travel. Besides, few work considers the hierarchical product taxonomy and the users attention to products explicitly. The fact is that users tend to express their attention to the semantic attributes of products such as color and style as the dialog goes on. Towards this end, in this work, we present a hierarchical User attention-guided Multimodal Dialog system, named UMD for short. UMD leverages a bidirectional Recurrent Neural Network to model the ongoing dialog between users and chatbots at a high level; As to the low level, the multimodal encoder and decoder are capable of encoding multimodal utterances and generating multimodal responses, respectively. The multimodal encoder learns the visual presentation of images with the help of a taxonomy-attribute combined tree, and then the visual features interact with textual features through an attention mechanism; whereas the multimodal decoder selects the required visual images and generates textual responses according to the dialog history. To evaluate our proposed model, we conduct extensive experiments on a public multimodal dialog dataset in the retailing domain. Experimental results demonstrate that our model outperforms the existing state-of-the-art methods by integrating the multimodal utterances and encoding the visual features based on the users attribute-level attention.

Keywords:  
User Attention-guided
Multimodal
Dialog Systems
Deep Learning
Machine Learning

Author(s) Name:  Chen Cui , Wenjie Wang , Xuemeng Song , Minlie Huang , Xin-Shun Xu , Liqiang Nie

Journal name:  

Conferrence name:  SIGIR-19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval

Publisher name:  ACM

DOI:  10.1145/3331184.3331226

Volume Information:  Pages 445–454