Abstract
The structural change toward the digital transformation of online sales elevates the importance of parallel processingtechniques in recommender systems, particularly in the pandemic and post-pandemic era. Matrix factorization (MF) is a popular andscalable approach in collaborative filtering (CF) to predict user preferences in recommender systems. Researchers apply StochasticGradient Descent (SGD) as one of the most famous optimization techniques for MF. Paralleling SGD methods help address big datachallenges due to the wide range of products and the sparsity in user ratings. However, these methods’ convergence rate and accuracyare affected by the dependency between the user and item latent factors, specifically in large-scale problems. Besides, the performance is sensitive to the applied learning rates. This article proposes a new parallel method to remove dependencies and boost speed-up by using fractional calculus to improve accuracy and convergence rate. We also apply adaptive learning rates to enhance the performance of our proposed method. The proposed method is based on Compute Unified Device Architecture (CUDA) platform. We evaluate the performance of our proposed method using real-world data and compare the results with the close baselines. The results show that our method can obtain high accuracy and convergence rate in addition to high parallelism.
Original language | English |
---|---|
Article number | 35 |
Pages (from-to) | 470-483 |
Number of pages | 14 |
Journal | IEEE Transactions on Parallel and Distributed Systems |
Volume | 35 |
Issue number | 3 |
Early online date | 23 Jun 2022 |
DOIs | |
Publication status | Published - 1 Mar 2024 |
Keywords
- Collaborative filtering
- Collaborative Filtering
- Convergence
- Graphics processing units
- Parallel Matrix Factorization
- Recommender Systems
- Recommender systems
- Sparse matrices
- Standards
- Stochastic processes