Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

3DSSCN: A Sentiment Analysis Model for Short Video Based on 3D-Dense Network

Version 1 : Received: 6 December 2022 / Approved: 7 December 2022 / Online: 7 December 2022 (11:57:32 CET)

How to cite: silva, M.; Kruger, H.; Vieira, M.; Stellato, C. 3DSSCN: A Sentiment Analysis Model for Short Video Based on 3D-Dense Network. Preprints 2022, 2022120132. https://doi.org/10.20944/preprints202212.0132.v1 silva, M.; Kruger, H.; Vieira, M.; Stellato, C. 3DSSCN: A Sentiment Analysis Model for Short Video Based on 3D-Dense Network. Preprints 2022, 2022120132. https://doi.org/10.20944/preprints202212.0132.v1

Abstract

In recent years, with the development of social media, people are more and more inclined to upload text, pictures and videos on the platform to express their personal emotions, thus the number of short videos is increasing and becoming the first choice for people to socialize. Unlike the traditional way, people can convey their personal emotions and opinions through media other than words, such as video images, etc. for external information. Therefore, the expression and analysis of emotions is not only through text, but also through the analysis of emotional needs in images and videos, and the research scholars have customized products for individual users. Compared with pure text content, video information can more intuitively express users' happiness, anger and sorrow, thus short video-related applications have gained more and more popularity among Internet users in recent years. However, not all short videos on social networking sites can accurately express users' emotions, and related text information can more accurately assist sentiment analysis and thus improve accuracy. However, short video sentiment analysis based on video frame images is inaccurate in some scenarios, such as when expressing tears of joy, the sentiment expressed by the user's facial expression and voice are different, which will cause errors in the analysis of sentiment. As a result, researchers began to consider multimodal sentiment analysis to reduce the impact of the above scenarios on short video sentiment analysis. This paper focuses on proposing a sentiment analysis method for short videos. We first propose a residual attention model to make full use of the information in audio to classify the emotions contained in them. Then the text information in the dataset is classified by feature extraction. The key to extract features from text information is not only to retain the semantic information of the text, but also to explore the potential emotional information in the text, so as to ensure the integrity of the text information features. The experiments show that the sentiment analysis model proposed in this paper is more superior than the baselines.

Keywords

Short video; Sentiment Analysis; Feature; 3D Dense Net; 3D Residual Network

Subject

Computer Science and Mathematics, Information Systems

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.