Show simple item record

dc.contributor.advisorAykanat, Cevdet
dc.contributor.authorSelvitopi, Oğuz
dc.date.accessioned2016-08-26T11:19:13Z
dc.date.available2016-08-26T11:19:13Z
dc.date.copyright2016-07
dc.date.issued2016-08
dc.date.submitted2016-08-16
dc.identifier.urihttp://hdl.handle.net/11693/32167
dc.descriptionCataloged from PDF version of article.en_US
dc.descriptionThesis (Ph.D.): Bilkent University, Department of Computer Engineering, İhsan Doğramacı Bilkent University, 2016.en_US
dc.descriptionIncludes bibliographical references (leaves 140-150).en_US
dc.description.abstractParallelization of sparse kernels and operations on large-scale distributed memory systems remains as a major challenge due to ever-increasing scale of modern high performance computing systems and multiple con icting factors that affect the parallel performance. The low computational density and high memory footprint of sparse operations add to these challenges by implying more stressed communication bottlenecks and make fast and effcient parallelization models and methods imperative for scalable performance. Sparse operations are usually performed with structures related to sparse matrices and matrices are partitioned prior to the execution for distributing computations among processors. Although the literature is rich in this aspect, it still lacks the techniques that embrace multiple factors affecting communication performance in a complete and just manner. In this thesis, we investigate models and methods for intelligent partitioning of sparse matrices that strive for achieving a more correct approximation of the communication performance. To improve the communication performance of parallel sparse operations, we mainly focus on reducing the latency bottlenecks, which stand as a major component in the overall communication cost. Besides these, our approaches consider already adopted communication cost metrics in the literature as well and aim to address as many cost metrics as possible. We propose one-phase and two-phase partitioning models to reduce the latency cost in one-dimensional (1D) and two-dimensional (2D) sparse matrix partitioning, respectively. The model for 1D partitioning relies on the commonly adopted recursive bipartitioning framework and it uses novel structures to capture the relations that incur latency. The models for 2D partitioning aim to improve the performance of solvers for nonsymmetric linear systems by using different partitions for the vectors in the solver and uses that exibility to exploit the latency cost. Our findings indicate that the latency costs should deffinitely be considered in order to achieve scalable performance on distributed memory systems.en_US
dc.description.statementofresponsibilityby Oğuz Selvitopi.en_US
dc.format.extentxv, 150 leaves : charts (some color)en_US
dc.language.isoEnglishen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.subjectParallel computingen_US
dc.subjectDistributed memory systemsen_US
dc.subjectCombinatorial scientific computingen_US
dc.subjectSparse matricesen_US
dc.subjectLoad balancingen_US
dc.subjectCommunication bottlenecksen_US
dc.subjectGraph partitioningen_US
dc.subjectHypergraph partitioningen_US
dc.titleLatency-centric models and methods for scaling sparse operationsen_US
dc.title.alternativeSeyrek işlemlerin ölçeklenebilmesi için gecikim-merkezli model ve yöntemleren_US
dc.typeThesisen_US
dc.departmentDepartment of Computer Engineeringen_US
dc.publisherBilkent Universityen_US
dc.description.degreePh.D.en_US
dc.identifier.itemidB153932
dc.embargo.release2018-08-01


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record