Scaling stratified stochastic gradient descent for distributed matrix completion
buir.contributor.author | Abubaker , Nabil | |
buir.contributor.author | Aykanat, Cevdet | |
buir.contributor.orcid | Abubaker, Nabil|0000-0002-5060-3059 | |
buir.contributor.orcid | Aykanat, Cevdet|0000-0002-4559-1321 | |
dc.citation.epage | 10615 | en_US |
dc.citation.issueNumber | 10 | |
dc.citation.spage | 10603 | |
dc.citation.volumeNumber | 35 | |
dc.contributor.author | Abubaker, Nabil | |
dc.contributor.author | Karsavuran, M. O. | |
dc.contributor.author | Aykanat, Cevdet | |
dc.date.accessioned | 2024-03-18T14:10:02Z | |
dc.date.available | 2024-03-18T14:10:02Z | |
dc.date.issued | 2023-10-01 | |
dc.department | Department of Computer Engineering | |
dc.description.abstract | Stratified SGD (SSGD) is the primary approach for achieving serializable parallel SGD for matrix completion. State-of-the-art parallelizations of SSGD fail to scale due to large communication overhead. During an SGD epoch, these methods send data proportional to one of the dimensions of the rating matrix. We propose a framework for scalable SSGD through significantly reducing the communication overhead via exchanging point-to-point messages utilizing the sparsity of the rating matrix. We provide formulas to represent the essential communication for correctly performing parallel SSGD and we propose a dynamic programming algorithm for efficiently computing them to establish the point-to-point message schedules. This scheme, however, significantly increases the number of messages sent by a processor per epoch from O(K) to (K2) for a K-processor system which might limit the scalability. To remedy this, we propose a Hold-and-Combine strategy to limit the upper-bound on the number of messages sent per processor to O(KlgK). We also propose a hypergraph partitioning model that correctly encapsulates reducing the communication volume. Experimental results show that the framework successfully achieves a scalable distributed SSGD through significantly reducing the communication overhead. Our code is publicly available at: github.com/nfabubaker/CESSGD | |
dc.description.provenance | Made available in DSpace on 2024-03-18T14:10:02Z (GMT). No. of bitstreams: 1 Scaling_stratified_stochastic_gradient_descent_for_distributed_matrix_completion.pdf: 1695216 bytes, checksum: f7959910441a0a141f8f04b2da21d575 (MD5) Previous issue date: 2023-10-01 | en |
dc.identifier.doi | 10.1109/TKDE.2023.3253791 | en_US |
dc.identifier.eissn | 1558-2191 | en_US |
dc.identifier.issn | 1041-4347 | en_US |
dc.identifier.uri | https://hdl.handle.net/11693/114918 | en_US |
dc.language.iso | English | en_US |
dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
dc.relation.isversionof | https://dx.doi.org/10.1109/TKDE.2023.3253791 | |
dc.source.title | IEEE Transactions on Knowledge and Data Engineering | |
dc.subject | Bandwidth cost | |
dc.subject | Combinatorial algorithms | |
dc.subject | Communication cost minimization | |
dc.subject | Collaborative filtering | |
dc.subject | HPC | |
dc.subject | Hypergraph partitioning | |
dc.subject | Latency cost | |
dc.subject | Matrix completion | |
dc.subject | Recommender systems | |
dc.subject | SGD | |
dc.title | Scaling stratified stochastic gradient descent for distributed matrix completion | |
dc.type | Article |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Scaling_stratified_stochastic_gradient_descent_for_distributed_matrix_completion.pdf
- Size:
- 1.58 MB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 2.01 KB
- Format:
- Item-specific license agreed upon to submission
- Description: