Publication Type

Conference Proceeding Article

Version

publishedVersion

Publication Date

6-2022

Abstract

Facebook has recently launched the hateful meme detection challenge, which garnered much attention in academic and industry research communities. Researchers have proposed multimodal deep learning classification methods to perform hateful meme detection. While the proposed methods have yielded promising results, these classification methods are mostly supervised and heavily rely on labeled data that are not always available in the real-world setting. Therefore, this paper explores and aims to perform hateful meme detection in a zero-shot setting. Working towards this goal, we propose Target-Aware Multimodal Enhancement (TAME), which is a novel deep generative framework that can improve existing hateful meme classification models’ performance in detecting unseen types of hateful memes. We conduct extensive experiments on the Facebook hateful meme dataset, and the results show that TAME can significantly improve the state-of-the-art hateful meme classification methods’ performance in seen and unseen settings.

Keywords

Hateful memes, Multimodal, Social media mining

Discipline

Artificial Intelligence and Robotics | Databases and Information Systems | Graphics and Human Computer Interfaces

Research Areas

Data Science and Engineering

Publication

Proceedings of the WebSci '22: 14th ACM Web Science Conference, Barcelona Spain, June 26 - 29

First Page

382

Last Page

389

ISBN

9781450391917

Identifier

10.1145/3501247.3531557

Publisher

ACM

City or Country

New York

Additional URL

https://doi.org/10.1145/3501247.3531557

Share

COinS