2:40 PM - 3:00 PM
[3K4-IS-2a-04] MORE: Modality-Embracing Contrastive Learning for Multimodal recommendation
Keywords:Recommendation System, Contrastive Learning, Multi-Modal
Multimodal recommendation helps users find their items of interest by utilizing items’ multimodal features, such as visual and textual modalities, in addition to interaction information with alleviating information overload problems. Although significant progress has been made on this challenge, existing research remains limitation of the modality embracing. Specifically, current research focuses on collaborative filtering signals, while the information included in the content (modality information) is not effectively represented in the constructed model. In this context, MONET proposes a well-designed Graph Convolutional Networks (GCNs) and achieves state-of-the-art performance for multimodal recommendations. However, it can be pointed out that employing specific GCNs architectures is insufficient to enhance retention rate of modalities. To address this limitation, we propose a simple yet effective model named Modality-Embracing COntRastive LEarning (MORE), which leveraging one of the self-supervised methods contrastive learning to synchronize modality information, thus enhancing the final embedding and the quality of recommendations. Our comprehensive experiments across two public datasets validate the enhanced performance of the MORE model.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.