Deep learning with extended exponential linear unit (DELU)

buir.contributor.authorÇatalbaş, Burak
buir.contributor.authorMorgül, Ömer
buir.contributor.orcidÇatalbaş, Burak|0000-0001-6235-3766
dc.citation.epage22724en_US
dc.citation.issueNumber30
dc.citation.spage22705
dc.citation.volumeNumber35
dc.contributor.authorÇatalbaş, Burak
dc.contributor.authorMorgül, Ömer
dc.date.accessioned2024-03-18T09:27:06Z
dc.date.available2024-03-18T09:27:06Z
dc.date.issued2023-08-16
dc.departmentDepartment of Electrical and Electronics Engineering
dc.description.abstractActivation functions are crucial parts of artificial neural networks. From the first perceptron created artificially up to today, many functions are proposed. Some of them are currently in common use, such as Rectified Linear Unit (ReLU) and Exponential Linear Unit (ELU) and other ReLU variants. In this article we propose a novel activation function, called ExtendeD Exponential Linear Unit (DELU). After its introduction and presenting its basic properties, by making various simulations with different datasets and architectures, we show that it may perform better than other activation functions in certain cases. While also inheriting most of the good properties of ReLU and ELU, DELU offers an increase of success in comparison with them by slowing the alignment of neurons in early stages of training process. In experiments, DELU performed better than other activation functions in general, for Fashion MNIST, CIFAR-10 and CIFAR-100 classification tasks with different sized Residual Neural Networks (ResNet). Specifically, DELU managed to reduce the error rate by sufficiently high confidence levels in CIFAR datasets in comparison with ReLU and ELU networks. In addition, DELU is compared in an image segmentation example as well. Also, compatibility of DELU is tested with different initializations, and statistical methods are employed to verify these success rates by using Z-score analysis, which may be considered as a different view of success assessment in neural networks.
dc.description.provenanceMade available in DSpace on 2024-03-18T09:27:06Z (GMT). No. of bitstreams: 1 Deep_learning_with_extended_exponential_linear_unit_(DELU).pdf: 2238469 bytes, checksum: 7500a1f65ce2a25918714e2a2b35b2de (MD5) Previous issue date: 2023-10en
dc.identifier.doi10.1007/s00521-023-08932-z
dc.identifier.eissn1433-3058
dc.identifier.issn0941-0643
dc.identifier.urihttps://hdl.handle.net/11693/114873
dc.language.isoen
dc.publisherSpringer
dc.relation.isversionofhttps://doi.org/10.1007/s00521-023-08932-z
dc.rightsCC BY 4.0 Deed (Attribution 4.0 International)
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.source.titleNeural Computing and Applications
dc.subjectActivation functions
dc.subjectArtificial neural networks
dc.subjectClassification
dc.subjectImage segmentation
dc.titleDeep learning with extended exponential linear unit (DELU)
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Deep_learning_with_extended_exponential_linear_unit_(DELU).pdf
Size:
2.13 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.01 KB
Format:
Item-specific license agreed upon to submission
Description: