Processing Multimodal Information

Challenges and Solutions for Multimodal Sentiment Analysis and Hate Speech Detection

authored by
Sherzod Hakimov, Gullal S. Cheema, Ralph Ewerth
Abstract

This chapter explores the challenges and solutions for processing multimodal information, specifically in the context of multimodal sentiment analysis and hate speech detection. The increasing amount of multimodal data, such as text, images and videos, presents unique challenges for machine learning algorithms. These challenges include the integration and fusion of information from multiple modalities to acquire the overall context. In this chapter, first, we present an overview of recent developments on multimodal learning techniques in the context of sentiment and hate speech detection; second, we present a multimodal model that combines different visual aspects and features for multimodal sentiment detection; and third, we present a multi-task multimodal model for misogyny detection in multimodal memes.

Organisation(s)
L3S Research Centre
External Organisation(s)
University of Potsdam
German National Library of Science and Technology (TIB)
Type
Contribution to book/anthology
Pages
71-94
No. of pages
24
Publication date
2025
Publication status
Published
Peer reviewed
Yes
ASJC Scopus subject areas
General Computer Science, General Social Sciences
Sustainable Development Goals
SDG 5 - Gender Equality
Electronic version(s)
https://doi.org/10.1007/978-3-031-64451-1_4 (Access: Open)