Multimodal Meme Dataset (MultiOFF) for Identifying Offensive Content in Image and Text

Main Authors: Shardul Suryawanshi, Bharathi Raja Chakravarthi, Mihael Arcan, Paul Buitelaar
Format: Proceeding Journal
Bahasa: eng
Terbitan: , 2020
Subjects:
Online Access: https://zenodo.org/record/3899868
ctrlnum 3899868
fullrecord <?xml version="1.0"?> <dc schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd"><creator>Shardul Suryawanshi</creator><creator>Bharathi Raja Chakravarthi</creator><creator>Mihael Arcan</creator><creator>Paul Buitelaar</creator><date>2020-05-11</date><description>A meme is a form of media that spreads an idea or emotion across the internet. As posting meme has become a new form of communication of the web, due to the multimodal nature of memes, postings of hateful memes or related events like trolling, cyberbullying are increasing day by day. Hate speech, offensive content and aggression content detection have been extensively explored in a single modality such as text or image. However, combining two modalities to detect offensive content is still a developing area. Memes make it even more challenging since they express humour and sarcasm in an implicit way, because of which the meme may not be offensive if we only consider the text or the image. Therefore, it is necessary to combine both modalities to identify whether a given meme is offensive or not. Since there was no publicly available dataset for multimodal offensive meme content detection, we leveraged the memes related to the 2016 U.S. presidential election and created the MultiOFF multimodal meme dataset for offensive content detection dataset. We subsequently developed a classifier for this task using the MultiOFF dataset. We use an early fusion technique to combine the image and text modality and compare it with a text- and an image-only baseline to investigate its effectiveness. Our results show improvements in terms of Precision, Recall, and F-Score. The code and dataset for this paper is published in https://github.com/bharathichezhiyan/ Multimodal-Meme-Classification-Identifying-Offensive-Content-in-Image-and-Text</description><identifier>https://zenodo.org/record/3899868</identifier><identifier>10.5281/zenodo.3899868</identifier><identifier>oai:zenodo.org:3899868</identifier><language>eng</language><relation>doi:10.5281/zenodo.3899867</relation><rights>info:eu-repo/semantics/openAccess</rights><rights>https://creativecommons.org/licenses/by/4.0/legalcode</rights><subject>multimodal data, classification, memes, offensive content, opinion mining</subject><title>Multimodal Meme Dataset (MultiOFF) for Identifying Offensive Content in Image and Text</title><type>Journal:Proceeding</type><type>Journal:Proceeding</type><recordID>3899868</recordID></dc>
language eng
format Journal:Proceeding
Journal
Journal:Journal
author Shardul Suryawanshi
Bharathi Raja Chakravarthi
Mihael Arcan
Paul Buitelaar
title Multimodal Meme Dataset (MultiOFF) for Identifying Offensive Content in Image and Text
publishDate 2020
topic multimodal data
classification
memes
offensive content
opinion mining
url https://zenodo.org/record/3899868
contents A meme is a form of media that spreads an idea or emotion across the internet. As posting meme has become a new form of communication of the web, due to the multimodal nature of memes, postings of hateful memes or related events like trolling, cyberbullying are increasing day by day. Hate speech, offensive content and aggression content detection have been extensively explored in a single modality such as text or image. However, combining two modalities to detect offensive content is still a developing area. Memes make it even more challenging since they express humour and sarcasm in an implicit way, because of which the meme may not be offensive if we only consider the text or the image. Therefore, it is necessary to combine both modalities to identify whether a given meme is offensive or not. Since there was no publicly available dataset for multimodal offensive meme content detection, we leveraged the memes related to the 2016 U.S. presidential election and created the MultiOFF multimodal meme dataset for offensive content detection dataset. We subsequently developed a classifier for this task using the MultiOFF dataset. We use an early fusion technique to combine the image and text modality and compare it with a text- and an image-only baseline to investigate its effectiveness. Our results show improvements in terms of Precision, Recall, and F-Score. The code and dataset for this paper is published in https://github.com/bharathichezhiyan/ Multimodal-Meme-Classification-Identifying-Offensive-Content-in-Image-and-Text
id IOS16997.3899868
institution ZAIN Publications
institution_id 7213
institution_type library:special
library
library Cognizance Journal of Multidisciplinary Studies
library_id 5267
collection Cognizance Journal of Multidisciplinary Studies
repository_id 16997
subject_area Multidisciplinary
city Stockholm
province INTERNASIONAL
shared_to_ipusnas_str 1
repoId IOS16997
first_indexed 2022-06-06T05:29:11Z
last_indexed 2022-06-06T05:29:11Z
recordtype dc
_version_ 1734905432482250752
score 17.608969