Amazing technological breakthrough possible @S-Logix pro@slogix.in

Office Address

  • #5, First Floor, 4th Street Dr. Subbarayan Nagar Kodambakkam, Chennai-600 024 Landmark : Samiyar Madam
  • pro@slogix.in
  • +91- 81240 01111

Social List

Rumor Detection on Social Media: A Multi-view Model Using Self-attention Mechanism - 2019

Rumor Detection On Social Media: A Multi-View Model Using Self-Attention Mechanism

Research Paper on Rumor Detection On Social Media: A Multi-View Model Using Self-Attention Mechanism

Research Area:  Machine Learning

Abstract:

With the unprecedented prevalence of social media, rumor detection has become increasingly important since it can prevent misinformation from spreading in public. Traditional approaches extract features from the source tweet, the replies, the user profiles as well as the propagation path of a rumor event. However, these approaches do not take the sentiment view of the users into account. The conflicting affirmative or denial stances of users can provide crucial clues for rumor detection. Besides, the existing work attaches the same importance to all the words in the source tweet, but actually, these words are not equally informative. To address these problems, we propose a simple but effective multi-view deep learning model that is supposed to excavate stances of users and assign weights for different words. Experimental results on a social-media based dataset reveal that the multi-view model we proposed is useful, and achieves the state-of-the-art performance measuring the accuracy of automatic rumor detection. Our three-view model achieves 95.6% accuracy and our four-view model using BERT as a view also reaches an improvement of detection accuracy.

Keywords:  
Rumor Detection
Social Media
Self-Attention Mechanism
Machine Learning
Deep Learning

Author(s) Name:  Yue GengZheng Lin, Peng FuWeiping Wang

Journal name:  

Conferrence name:  International Conference on Computational Science

Publisher name:  Springer

DOI:  https://doi.org/10.1007/978-3-030-22734-0_25

Volume Information: