Improving Rumor Detection by Image Captioning and Multi-Cell Bi-RNN With Self-Attention in Social Networks

Improving Rumor Detection by Image Captioning and Multi-Cell Bi-RNN With Self-Attention in Social Networks

Jenq-Haur Wang, Chin-Wei Huang, Mehdi Norouzi
Copyright: © 2022 |Pages: 17
DOI: 10.4018/IJDWM.313189
Article PDF Download
Open access articles are freely available for download

Abstract

User-generated contents in social media are not verified before being posted. They could bring many problems if they were misused. Among various types of rumors, the authors focus on the type in which there's mismatch between images and their surrounding texts. They can be detected by multimodal feature fusion in RNNs with attention mechanism, but the relations between images and texts are not well-addressed. In this paper, the authors propose to improve rumor detection by image captioning and RNNs with self-attention. Firstly, they utilize the idea of image captioning to translate images into the corresponding text descriptions. Secondly, these caption words are represented by word embedding models and aggregated with surrounding texts using early fusion. Finally, multi-cell bi-directional RNNs with self-attention are used to learn important features to identify rumors. From the experimental results, the best F-measure of 0.882 can be obtained, which shows the potential of our proposed approach to rumor detection. Further investigation is needed for data in larger scale.
Article Preview
Top

1. Introduction

With the rapid development of information and communication technology, people can easily share their opinions and get the latest news from social network platforms. Since these user-generated contents are not verified before being posted, people cannot tell if they are real or false. Thus, the unverified false messages are considered as rumors since they are annoying in various aspects of our daily lives. For example, people could receive messages that masquerade as being sent from the government or companies, asking to provide personal information. People can only tell if these messages are real or false by checking if the text contents are relevant or not. To distinguish real messages from false ones, rumor detection has been an increasingly important research topic in social media. Nowadays, third-party fact-checking services such as FactCheck.org and Snopes.com have been used for message verification. These fact-checking services usually require manual labeling which needs lots of human efforts. In the face of rapid information dissemination, these services alone cannot be effectively provided in time.

There could be numerous ways to disguise the unverified false messages or rumors as real messages. Since it’s very common to post images and texts in the same post in social media, in this paper we define rumors as the message type when there’s mismatch between multimedia contents and the surrounding texts. Our research problem for rumor detection is defined as: given a social media post with images and their corresponding surrounding texts, we want to determine if there’s a mismatch between semantic information in the images and the surrounding texts. In recent research, deep learning methods are widely used to construct the model and to learn features for rumor detection. For example, as the baseline in our experiments, the method proposed by Jin et. al (2017) utilized a RNN with attention mechanism (att-RNN) to fuse multimodal features, including texts and images. They achieved an accuracy of 68.2% for the Twitter dataset in MediaEval task. The visual neuron attention mechanism gives each neuron different weights for different words. However, the relations between image visual features and text features are not well-addressed. In order to better address the relations between images and the surrounding texts, in this paper, we propose to improve rumor detection by image captioning and multi-cell RNNs with self-attention. Firstly, it’s very important if we can aggregate multimodal contents as an effective feature to tell the mismatch between texts and images. Instead of simply adjusting weights of different visual features by the attention mechanism from the RNN results of texts, we first translate images into the most relevant caption words as a more coherent way of feature representation. This helps to closely connect the semantic meanings of images and texts. Secondly, we design a novel type of multi-cell bidirectional RNNs which combine self-attention mechanism for identifying more important features from different sources. Finally, different feature fusion approaches are used to improve the performance of rumor detection.

The main contributions of this paper are as follows: Firstly, we propose a novel multimodal feature fusion approach to rumor detection based on image captioning model that represents image semantics in textual descriptions. The sequence-to-sequence model can extract more meaningful descriptions from images than simple convolutional approaches. To the best of our knowledge, our proposed method is the first to apply image captioning in extracting image semantics for rumor detection. Secondly, instead of one single layer which might not fully capture the relations among words, we design a novel way of stacking bidirectional RNNs called Multi-cell Bi-RNN, which adds more cells in each individual direction of forward and backward passes to learn more deeply in each neuron. A better performance can be obtained than the baseline model. This shows the potential of our proposed approach to rumor detection.

The rest of the paper is structured as follows: In Sec. 2, we provide a review of related previous research works. Then the proposed method is presented and discussed in Sec.3. In Sec.4, we show the experimental results and compare with existing methods. Finally, we conclude the paper in Sec.5.

Complete Article List

Search this Journal:
Reset
Volume 20: 1 Issue (2024)
Volume 19: 6 Issues (2023)
Volume 18: 4 Issues (2022): 2 Released, 2 Forthcoming
Volume 17: 4 Issues (2021)
Volume 16: 4 Issues (2020)
Volume 15: 4 Issues (2019)
Volume 14: 4 Issues (2018)
Volume 13: 4 Issues (2017)
Volume 12: 4 Issues (2016)
Volume 11: 4 Issues (2015)
Volume 10: 4 Issues (2014)
Volume 9: 4 Issues (2013)
Volume 8: 4 Issues (2012)
Volume 7: 4 Issues (2011)
Volume 6: 4 Issues (2010)
Volume 5: 4 Issues (2009)
Volume 4: 4 Issues (2008)
Volume 3: 4 Issues (2007)
Volume 2: 4 Issues (2006)
Volume 1: 4 Issues (2005)
View Complete Journal Contents Listing