Hardware acceleration for Swin Transformers at the edge

buir.advisorGüdükbay, Uğur
dc.contributor.authorEsergün, Yunus
dc.date.accessioned2024-06-06T10:55:16Z
dc.date.available2024-06-06T10:55:16Z
dc.date.copyright2024-05
dc.date.issued2024-05
dc.date.submitted2024-06-05
dc.descriptionCataloged from PDF version of article.
dc.descriptionThesis (Master's): Bilkent University, Department of Computer Engineering, İhsan Doğramacı Bilkent University, 2024.
dc.descriptionIncludes bibliographical references (leaves 46-51).
dc.description.abstractWhile deep learning models have greatly enhanced visual processing abilities, their implementation in edge environments with limited resources can be challenging due to their high energy consumption and computational requirements. Swin Transformer is a prominent mechanism in computer vision that differs from traditional convolutional approaches. It adopts a hierarchical approach to interpreting images. A common strategy that improves the efficiency of deep learning algorithms during inference is clustering. Locality-Sensitive Hashing (LSH) is a mechanism that implements clustering and leverages the inherent redundancy within Transformers to identify and exploit computational similarities. This the-sis introduces a hardware accelerator for Swin Transformer implementation with LSH in edge computing settings. The main goal is to reduce energy consumption while improving performance with custom hardware components. Specifically, our custom hardware accelerator design utilizes LSH clustering in Swin Transformers to decrease the amount of computation required. We tested our accelerator with two different state-of-the-art datasets, namely, Imagenet-1K and CIFAR-100. Our results demonstrate that the hardware accelerator enhances the processing speed of the Swin Transformer when compared to GPU-based implementations. More specifically, our accelerator improves performance by 1.35x while reducing the power consumption to 5-6 Watts instead of 19 Watts in the baseline GPU setting. We observe these improvements with a negligible decrease in model accuracy of less than 1%, confirming the effectiveness of our hardware accelerator design in edge computing environments with limited resources.
dc.description.provenanceMade available in DSpace on 2024-06-06T10:55:16Z (GMT). No. of bitstreams: 1 B128438.pdf: 7787998 bytes, checksum: 8d9ad21e3e6cc6a6666cc1b0b812d1f5 (MD5) Previous issue date: 2024-05en
dc.description.statementofresponsibilityby Yunus Eserrgün
dc.embargo.release2024-12-05
dc.format.extentx, 51 leaves : illustrations, charts ; 30 cm.
dc.identifier.itemidB128438
dc.identifier.urihttps://hdl.handle.net/11693/115183
dc.language.isoEnglish
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectSwin Transformer
dc.subjectLSH
dc.subjectHardware accelerator
dc.subjectInference
dc.subjectGPU
dc.subjectFPGA
dc.subjectEdge
dc.subjectPower
dc.titleHardware acceleration for Swin Transformers at the edge
dc.title.alternativeUçta Swin Tabanlı Dönüştürücüler için donanım hızlandırması
dc.typeThesis
thesis.degree.disciplineComputer Engineering
thesis.degree.grantorBilkent University
thesis.degree.levelMaster's
thesis.degree.nameMS (Master of Science)

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
B128438.pdf
Size:
7.43 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.01 KB
Format:
Item-specific license agreed upon to submission
Description: