Improving the performance of 1D vertex parallel GNN training on distributed memory systems

buir.advisorAykanat, Cevdet
dc.contributor.authorTaşcı, Kutay
dc.date.accessioned2024-08-08T13:48:12Z
dc.date.available2024-08-08T13:48:12Z
dc.date.copyright2024-07
dc.date.issued2024-07
dc.date.submitted2024-08-02
dc.descriptionCataloged from PDF version of article.
dc.descriptionThesis (Master's): Bilkent University, Department of Computer Engineering, İhsan Doğramacı Bilkent University, 2024.
dc.descriptionIncludes bibliographical references (leaves 49-52).
dc.description.abstractGraph Neural Networks (GNNs) are pivotal for analyzing data within graphstructured domains such as social media, biological networks, and recommendation systems. Despite their advantages, scaling GNN training to large datasets in distributed settings poses significant challenges due to the complex task of managing computation and communication costs. The objective of this work is to scale 1D vertex-parallel GNN training on distributed memory systems via (i) twoconstraint partitioning formulation for better computational load balancing and (ii) overlapping communication with computation for reducing communication overhead. In the proposed two-constraint formulation, one constraint encodes the computational load balance during forward propagation, whereas the second constraint encodes the computational load balance during backward propagation. We propose three communication and computation overlapping methods that perform overlapping at three different levels. These methods were tested against traditional approaches using benchmark datasets, demonstrating improved training efficiency without altering the model structure. The outcomes indicate that multi-constraint graph partitioning and the integration of communication and computation overlapping schemes can significantly mitigate the challenges of distributed GNN training. The research concludes with recommendations for future work, including adapting these techniques to dynamic and more complex GNN architectures, promising further improvements in the efficiency and applicability of GNNs in real-world scenarios.
dc.description.provenanceSubmitted by Betül Özen (ozen@bilkent.edu.tr) on 2024-08-08T13:48:12Z No. of bitstreams: 1 B138237.pdf: 363621 bytes, checksum: b4ae43abed7422047837ac44ddfb2559 (MD5)en
dc.description.provenanceMade available in DSpace on 2024-08-08T13:48:12Z (GMT). No. of bitstreams: 1 B138237.pdf: 363621 bytes, checksum: b4ae43abed7422047837ac44ddfb2559 (MD5) Previous issue date: 2024-07en
dc.description.statementofresponsibilityby Kutay Taşcı
dc.format.extentx, 52 leaves : charts ; 30 cm.
dc.identifier.itemidB138237
dc.identifier.urihttps://hdl.handle.net/11693/115727
dc.language.isoEnglish
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectGraph neural networks
dc.subjectParallel and distributed memory systems
dc.subjectGraph partitioning
dc.subjectLoad balancing
dc.subjectOverlapping communication with computation
dc.titleImproving the performance of 1D vertex parallel GNN training on distributed memory systems
dc.title.alternativeDağıtık bellek sistemlerinde 1D düğüm paralel GNN eğitiminin performansının iyileştirilmesi
dc.typeThesis
thesis.degree.disciplineComputer Engineering
thesis.degree.grantorBilkent University
thesis.degree.levelMaster's
thesis.degree.nameMS (Master of Science)

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
B138237.pdf
Size:
355.1 KB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.1 KB
Format:
Item-specific license agreed upon to submission
Description: