Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Dictionary Learning for Personalized Multimodal Recommendation

Version 1 : Received: 28 June 2022 / Approved: 30 June 2022 / Online: 30 June 2022 (03:43:30 CEST)

How to cite: Liu, B.; Millett, H.; Rebola, B.L.; Svensen, P. Dictionary Learning for Personalized Multimodal Recommendation. Preprints 2022, 2022060412. https://doi.org/10.20944/preprints202206.0412.v1 Liu, B.; Millett, H.; Rebola, B.L.; Svensen, P. Dictionary Learning for Personalized Multimodal Recommendation. Preprints 2022, 2022060412. https://doi.org/10.20944/preprints202206.0412.v1

Abstract

In today’s Web 2.0 era, online social media has become an integral part of our lives. In the course of the information revolution, the form of information has undergone a radical change, from simple text information to today’s integrated video, image, text and audio, and there has also been a great change in the way of dissemination and access, as people nowadays do not just rely on traditional media to passively receive information, but more actively and selectively obtain information from social media. Therefore, it has become a great challenge for us to effectively utilize these massive and integrated multi-modal media information to form an effective system of retrieval, browsing, analysis and usage. Unlike movies and traditional long-form video content, micro-videos are usually short in length, between a few seconds and tens of seconds, which allows users to quickly browse different contents and make full use of the fragmented time in their lives, while users can also share their micro-videos to their friends or the public, forming a unique social way. Video contains rich multimodal information, and fusing information from multiple modalities in a video recommendation task can improve the accuracy of the video recommendation task.According to the micro-video recommendation task, a new combinatorial network model is proposed to combine the discrete features of each modality into the overall features of various modalities through the network, and then fuse the various modal features to obtain the overall video features, which will be used for recommendation. In order to verify the effectiveness of the algorithm proposed in this paper, experiments are conducted in the public dataset, and it is shown the effectiveness of our model.

Keywords

Dictionary learning, Recommender system, Personalized recommendation, Multimodal

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.