Memory efficient filtering algorithms for convolutional neural networks

buir.advisorMorgül, Ömer
dc.contributor.authorÇakır, Bahadır Alp
dc.date.accessioned2021-01-08T08:50:19Z
dc.date.available2021-01-08T08:50:19Z
dc.date.copyright2020-12
dc.date.issued2020-12
dc.date.submitted2021-01-07
dc.departmentDepartment of Electrical and Electronics Engineeringen_US
dc.descriptionCataloged from PDF version of article.en_US
dc.descriptionThesis (Master's): Bilkent University, Department of Electrical and Electronics Engineering, İhsan Doğramacı Bilkent University, 2020.en_US
dc.descriptionIncludes bibliographical references (leaves 51-56).en_US
dc.description.abstractDeployment of state of the art CNN architectures like Xception, ResNet and GoogleNet in resource limited devices is a big challenge. These architectures consist of many layers and millions of parameters. Moreover, they require billions of floating point operations to inference just an image. Therefore, memory space needed to store parameters and to execute them are the main constraints for efficient convolutional neural network architectures. In this thesis, we examine Winograd’s minimal filtering algorithms to reduce number of floating point operations performed in convolutional layers. We reduce the number of multiplications x2.25 times without any accuracy loss. Moreover, we investigate, sparse and quantized Winograd’s algorithms so that we can make conventional Winograd algorithms more memory efficient. We propose a linear quantization scheme to quantize weights of the networks more than 1-bit. We use ReLU activation function and Targeted Dropout which is a variant of Dropout to prune transformed inputs of Winograd algorithm. We binarize weights so that most arithmetic operations are converted to bit-wise operations. We conduct several experiments on CIFAR10 and CIFAR100 datasets and discuss the classification performances of both conventional and modified Winograd minimal filtering algorithms. We achieve less than 1.9% classification error with ReLU-ed Winograd CNN compared to conventional Winograd. We reduce memory requirements up to x32 times by binarizing weights of ReLU-ed Winograd CNN, and in return we incur around 2% accuracy loss. Lastly, for applications which are less tolerant to accuracy loss, rather than binarizing weights we quantize them to 2-bit, 4-bit and 8-bit. Our quantized ReLU-ed Winograd CNNs reach same accuracy levels as ReLU-ed Winograd CNN.en_US
dc.description.degreeM.S.en_US
dc.description.statementofresponsibilityby Bahadır Alp Çakıren_US
dc.format.extentxii, 56 leaves : some color, charts ; 30 cm.en_US
dc.identifier.itemidB130916
dc.identifier.urihttp://hdl.handle.net/11693/54874
dc.language.isoEnglishen_US
dc.publisherBilkent Universityen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.subjectWinograd’s minimal filtering algorithmsen_US
dc.subjectReLUen_US
dc.subjectTargeted dropouten_US
dc.subjectBinary weightsen_US
dc.subjectQuantized weightsen_US
dc.subjectMemory efficiencyen_US
dc.titleMemory efficient filtering algorithms for convolutional neural networksen_US
dc.title.alternativeEvrişimli yapay sinir ağları için bellek verimli filtreleme algoritmalarıen_US
dc.typeThesisen_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
10373488.pdf
Size:
1.22 MB
Format:
Adobe Portable Document Format
Description:
Full printable version
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: