Hardware acceleration for Swin Transformers at the edge
buir.advisor | Güdükbay, Uğur | |
dc.contributor.author | Esergün, Yunus | |
dc.date.accessioned | 2024-06-06T10:55:16Z | |
dc.date.available | 2024-06-06T10:55:16Z | |
dc.date.copyright | 2024-05 | |
dc.date.issued | 2024-05 | |
dc.date.submitted | 2024-06-05 | |
dc.description | Cataloged from PDF version of article. | |
dc.description | Thesis (Master's): Bilkent University, Department of Computer Engineering, İhsan Doğramacı Bilkent University, 2024. | |
dc.description | Includes bibliographical references (leaves 46-51). | |
dc.description.abstract | While deep learning models have greatly enhanced visual processing abilities, their implementation in edge environments with limited resources can be challenging due to their high energy consumption and computational requirements. Swin Transformer is a prominent mechanism in computer vision that differs from traditional convolutional approaches. It adopts a hierarchical approach to interpreting images. A common strategy that improves the efficiency of deep learning algorithms during inference is clustering. Locality-Sensitive Hashing (LSH) is a mechanism that implements clustering and leverages the inherent redundancy within Transformers to identify and exploit computational similarities. This the-sis introduces a hardware accelerator for Swin Transformer implementation with LSH in edge computing settings. The main goal is to reduce energy consumption while improving performance with custom hardware components. Specifically, our custom hardware accelerator design utilizes LSH clustering in Swin Transformers to decrease the amount of computation required. We tested our accelerator with two different state-of-the-art datasets, namely, Imagenet-1K and CIFAR-100. Our results demonstrate that the hardware accelerator enhances the processing speed of the Swin Transformer when compared to GPU-based implementations. More specifically, our accelerator improves performance by 1.35x while reducing the power consumption to 5-6 Watts instead of 19 Watts in the baseline GPU setting. We observe these improvements with a negligible decrease in model accuracy of less than 1%, confirming the effectiveness of our hardware accelerator design in edge computing environments with limited resources. | |
dc.description.provenance | Made available in DSpace on 2024-06-06T10:55:16Z (GMT). No. of bitstreams: 1 B128438.pdf: 7787998 bytes, checksum: 8d9ad21e3e6cc6a6666cc1b0b812d1f5 (MD5) Previous issue date: 2024-05 | en |
dc.description.statementofresponsibility | by Yunus Eserrgün | |
dc.embargo.release | 2024-12-05 | |
dc.format.extent | x, 51 leaves : illustrations, charts ; 30 cm. | |
dc.identifier.itemid | B128438 | |
dc.identifier.uri | https://hdl.handle.net/11693/115183 | |
dc.language.iso | English | |
dc.rights | info:eu-repo/semantics/openAccess | |
dc.subject | Swin Transformer | |
dc.subject | LSH | |
dc.subject | Hardware accelerator | |
dc.subject | Inference | |
dc.subject | GPU | |
dc.subject | FPGA | |
dc.subject | Edge | |
dc.subject | Power | |
dc.title | Hardware acceleration for Swin Transformers at the edge | |
dc.title.alternative | Uçta Swin Tabanlı Dönüştürücüler için donanım hızlandırması | |
dc.type | Thesis | |
thesis.degree.discipline | Computer Engineering | |
thesis.degree.grantor | Bilkent University | |
thesis.degree.level | Master's | |
thesis.degree.name | MS (Master of Science) |