Dataset and Models for Item Recommendation Using Multi-Modal User Interactions

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningpeer review

2 Downloads (Pure)

Abstract

While recommender systems with multi-modal item representations (image, audio, and text), have been widely explored, learning recommendations from multi-modal user interactions (e.g., clicks and speech) remains an open problem. We study the case of multi-modal user interactions in a setting where users engage with a service provider through multiple channels (website and call center). In such cases, incomplete modalities naturally occur, since not all users interact through all the available channels. To address these challenges, we publish a real-world dataset that allows progress in this under-researched area. We further present and benchmark various methods for leveraging multi-modal user interactions for item recommendations, and propose a novel approach that specifically deals with missing modalities by mapping user interactions to a common feature space. Our analysis reveals important interactions between the different modalities and that a frequently occurring modality can enhance learning from a less frequent one.

OriginalsprogEngelsk
TitelSIGIR 2024 - Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval
Antal sider10
ForlagAssociation for Computing Machinery, Inc.
Publikationsdato2024
Sider709-718
ISBN (Elektronisk)9798400704314
DOI
StatusUdgivet - 2024
Begivenhed47th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2024 - Washington, USA
Varighed: 14 jul. 202418 jul. 2024

Konference

Konference47th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2024
Land/OmrådeUSA
ByWashington
Periode14/07/202418/07/2024
SponsorACM SIGIR

Bibliografisk note

Publisher Copyright:
© 2024 Owner/Author.

Citationsformater