Pure Cold Start Recommendation by Learning on Stochastically Expanded Graphs

More Info
expand_more

Abstract

Recommender systems help users to find items they presumably like based on data collected on that user. Collaborative filtering is arguably the most common recommendation system technique. It uses collected ratings to compute similarities between users and recommends items based on those similarities. With that, a problem arises when there are few ratings available for a user, also called the cold start problem. An even stricter setting is where no ratings of a user are available at all, called the pure cold start problem. Graphs are a natural way to represent data in this domain and Graph Neural Networks (GNNs) have proven to achieve state-of-the-art accuracies in the domain of recommender systems over the past few years. In this work, we address this problem from a graph perspective. In the current literature, the problem is alleviated by turning to external information, such as demographic or social information. Although these methods work, they cannot be deployed in situations where this information is not available. However, our approach uses only rating data. As ratings are unavailable for pure cold start users, their connectivity to the graph is unknown. To overcome this, we use a stochastic attachment model to infer connectivity. This model is composed of a Bernoulli random variable with the corresponding probability value and edge weight for each existing user. In this work, we propose to learn the parameters of a graph convolution model through empirical risk minimization using the stochastic attachment model to attach users who previously attached to the graph and their ratings, and use it to predict the same for unseen users. Furthermore, we propose to learn the probability and weight values of the stochastic attachment model jointly with the parameters of the graph convolution model. We compare these methods to several graph based and non-graph baselines. Furthermore, we compare different methods such as graph filters, GCNNs, and kNN. The experiments show significant improvements of the graph approaches compared to the non-graph baselines. Furthermore, we show that training on stochastically expanded graphs improves accuracy compared to only testing on them by induction.