Multi-modal Adaptive Mixture of Experts for Cold-start Recommendation
Van Khang Nguyen (Vietnam National University Hanoi)
Duc Hoang Pham (Vietnam National University Hanoi)
Huy Son Nguyen (TU Delft - Multimedia Computing)
Cam Van Thi Nguyen (Vietnam National University Hanoi)
Hoang Quynh Le (Vietnam National University Hanoi)
Duc Trong Le (Vietnam National University Hanoi)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Recommendation systems have faced significant challenges in cold-start scenarios, where new items with a limited history of interaction need to be effectively recommended to users. Though multimodal data (e.g., images, text, audio, etc.) offer rich information to address this issue, existing approaches often employ simplistic integration methods such as concatenation, average pooling, or fixed weighting schemes, which fail to capture the complex relationships between modalities. Our study proposes a novel Mixture of Experts framework for multimodal cold-start recommendation (MAMEX), which dynamically leverages latent representation from different modalities. MAMEX utilizes modality-specific expert networks and introduces a learnable gating mechanism that adaptively weights the contribution of each modality based on its content characteristics. This approach enables MAMEX to emphasize the most informative modalities for each item while maintaining robustness when certain modalities are less relevant or missing. Extensive experiments on benchmark datasets show that MAMEX outperforms state-of-the-art models with superior accuracy and adaptability.
Files
File under embargo until 10-05-2026