Minimizing staleness and communication overhead in distributed SGD for collaborative filtering

buir.contributor.authorAbubaker, Nabil
buir.contributor.authorAykanat, Cevdet
buir.contributor.orcidAbubaker, Nabil|0000-0002-5060-3059
buir.contributor.orcidAykanat, Cevdet|0000-0002-4559-1321
dc.citation.epage2937en_US
dc.citation.issueNumber10
dc.citation.spage2925
dc.citation.volumeNumber72
dc.contributor.authorAbubaker, Nabil
dc.contributor.authorCaglayan, O.
dc.contributor.authorKarsavuran, M. O.
dc.contributor.authorAykanat, Cevdet
dc.date.accessioned2024-03-18T13:10:40Z
dc.date.available2024-03-18T13:10:40Z
dc.date.issued2023-09-06
dc.departmentDepartment of Computer Engineering
dc.description.abstractDistributed asynchronous stochastic gradient descent (ASGD) algorithms that approximate low-rank matrix factorizations for collaborative filtering perform one or more synchronizations per epoch where staleness is reduced with more synchronizations. However, high number of synchronizations would prohibit the scalability of the algorithm. We propose a parallel ASGD algorithm, η-PASGD, for efficiently handling η synchronizations per epoch in a scalable fashion. The proposed algorithm puts an upper limit of KK on η, for a KK-processor system, such that performing Kη=K synchronizations per epoch would eliminate the staleness completely. The rating data used in collaborative filtering are usually represented as sparse matrices. The sparsity allows for reduction in the staleness and communication overhead combinatorially via intelligently distributing the data to processors. We analyze the staleness and the total volume incurred during an epoch of η-PASGD. Following this analysis, we propose a hypergraph partitioning model to encapsulate reducing staleness and volume while minimizing the maximum number of synchronizations required for a stale-free SGD. This encapsulation is achieved with a novel cutsize metric that is realized via a new recursive-bipartitioning-based algorithm. Experiments on up to 512 processors show the importance of the proposed partitioning method in improving staleness, volume, RMSE and parallel runtime.
dc.description.provenanceMade available in DSpace on 2024-03-18T13:10:40Z (GMT). No. of bitstreams: 1 Minimizing_staleness_and_communication_overhead_in_distributed_SGD_for_collaborative_filtering.pdf: 1512925 bytes, checksum: 5decc956e863b67a585995e04c30e601 (MD5) Previous issue date: 2023-10-01en
dc.identifier.doi10.1109/TC.2023.3275107
dc.identifier.eissn1557-9956
dc.identifier.issn0018-9340
dc.identifier.urihttps://hdl.handle.net/11693/114904
dc.language.isoEnglish
dc.publisherIEEE Computer Society
dc.relation.isversionofhttps://dx.doi.org/10.1109/TC.2023.3275107
dc.source.titleIEEE Transactions on Computers
dc.subjectRecommender systems
dc.subjectCollaborative filtering
dc.subjectMatrix completion
dc.subjectDistributed-memory parallel stochastic gradient descent
dc.subjectCommunication-efficient algorithms,
dc.subjectMPI
dc.subjectHypergraph partitioning
dc.titleMinimizing staleness and communication overhead in distributed SGD for collaborative filtering
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Minimizing_staleness_and_communication_overhead_in_distributed_SGD_for_collaborative_filtering.pdf
Size:
1.42 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.01 KB
Format:
Item-specific license agreed upon to submission
Description: