Publication Type

Conference Proceeding Article

Version

publishedVersion

Publication Date

8-2023

Abstract

Recent studies have proposed models that yielded promising performance for the hateful meme classification task. Nevertheless, these proposed models do not generate interpretable explanations that uncover the underlying meaning and support the classification output. A major reason for the lack of explainable hateful meme methods is the absence of a hateful meme dataset that contains ground truth explanations for benchmarking or training. Intuitively, having such explanations can educate and assist content moderators in interpreting and removing flagged hateful memes. This paper address this research gap by introducing Hateful meme with Reasons Dataset (HatReD), which is a new multimodal hateful meme dataset annotated with the underlying hateful contextual reasons. We also define a new conditional generation task that aims to automatically generate underlying reasons to explain hateful memes and establish the baseline performance of state-of-the-art pre-trained language models on this task. We further demonstrate the usefulness of HatReD by analyzing the challenges of the new conditional generation task in explaining memes in seen and unseen domains. The dataset and benchmark models are made available here: https://github.com/Social-AI-Studio/HatRed

Keywords

Knowledge Representation and Reasoning, Natural Language Processing, Computer Vision

Discipline

Artificial Intelligence and Robotics | Databases and Information Systems

Publication

Proceedings of the 32nd International Joint Conference on Artificial Intelligence, IJCAI 2023: Macao, August 19-25

First Page

5995

Last Page

6003

ISBN

9781956792034

Identifier

10.24963/ijcai.2023/665

Publisher

AAAI Press

City or Country

Washington, DC

Additional URL

https://doi.org/10.24963/ijcai.2023/665

Share

COinS