Preprint Article Version 1 This version is not peer-reviewed

Fusion-Enhanced Multimodal Social Media Analysis

Version 1 : Received: 9 August 2024 / Approved: 12 August 2024 / Online: 12 August 2024 (10:14:37 CEST)

How to cite: Jackson, A.; Patel, R.; Taylor, E.; Anderson, M. Fusion-Enhanced Multimodal Social Media Analysis. Preprints 2024, 2024080779. https://doi.org/10.20944/preprints202408.0779.v1 Jackson, A.; Patel, R.; Taylor, E.; Anderson, M. Fusion-Enhanced Multimodal Social Media Analysis. Preprints 2024, 2024080779. https://doi.org/10.20944/preprints202408.0779.v1

Abstract

The analysis of social media content plays a crucial role in uncovering intricate user behaviors and trends across diverse digital platforms. Social media is inherently multimodal, incorporating texts, images, audios, and videos, each offering unique insights into user engagement and preferences. Traditional classification methods often focus narrowly on the most prominent modality, typically neglecting the synergistic potential of integrating multiple data types. To address this gap, we introduce the Unified Multimodal Classifier (UMC), a suite of streamlined models that adeptly harness and integrate these varied modalities. UMC leverages a novel architecture that combines a pooling layer with auxiliary learning tasks, facilitating the formation of a robust and shared feature space. This integration allows UMC to not only accommodate but also capitalize on the inherent diversity of social media data. The models are designed to be inherently flexible, adjusting to the availability of data modalities and maintaining high classification accuracy under varied conditions. In emotion classification scenarios, UMC has shown exceptional performance, significantly outperforming traditional methods by effectively synthesizing information across modalities. Its robustness is highlighted in its ability to deliver consistent results, even in the absence of one or more modalities. The simplicity and efficiency of UMC make it a potent tool for social media analytics, enabling it to achieve higher accuracies and maintain effectiveness even with limited data inputs. This adaptive capability ensures that UMC can be deployed in real-world applications where data incompleteness is common, thus broadening the scope of its applicability in various analytical contexts.

Keywords

multimodal classification; social media analysis; emotion recognition

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.