Deep learning with extended exponential linear unit (DELU)
buir.contributor.author | Çatalbaş, Burak | |
buir.contributor.author | Morgül, Ömer | |
buir.contributor.orcid | Çatalbaş, Burak|0000-0001-6235-3766 | |
dc.citation.epage | 22724 | en_US |
dc.citation.issueNumber | 30 | |
dc.citation.spage | 22705 | |
dc.citation.volumeNumber | 35 | |
dc.contributor.author | Çatalbaş, Burak | |
dc.contributor.author | Morgül, Ömer | |
dc.date.accessioned | 2024-03-18T09:27:06Z | |
dc.date.available | 2024-03-18T09:27:06Z | |
dc.date.issued | 2023-08-16 | |
dc.department | Department of Electrical and Electronics Engineering | |
dc.description.abstract | Activation functions are crucial parts of artificial neural networks. From the first perceptron created artificially up to today, many functions are proposed. Some of them are currently in common use, such as Rectified Linear Unit (ReLU) and Exponential Linear Unit (ELU) and other ReLU variants. In this article we propose a novel activation function, called ExtendeD Exponential Linear Unit (DELU). After its introduction and presenting its basic properties, by making various simulations with different datasets and architectures, we show that it may perform better than other activation functions in certain cases. While also inheriting most of the good properties of ReLU and ELU, DELU offers an increase of success in comparison with them by slowing the alignment of neurons in early stages of training process. In experiments, DELU performed better than other activation functions in general, for Fashion MNIST, CIFAR-10 and CIFAR-100 classification tasks with different sized Residual Neural Networks (ResNet). Specifically, DELU managed to reduce the error rate by sufficiently high confidence levels in CIFAR datasets in comparison with ReLU and ELU networks. In addition, DELU is compared in an image segmentation example as well. Also, compatibility of DELU is tested with different initializations, and statistical methods are employed to verify these success rates by using Z-score analysis, which may be considered as a different view of success assessment in neural networks. | |
dc.description.provenance | Made available in DSpace on 2024-03-18T09:27:06Z (GMT). No. of bitstreams: 1 Deep_learning_with_extended_exponential_linear_unit_(DELU).pdf: 2238469 bytes, checksum: 7500a1f65ce2a25918714e2a2b35b2de (MD5) Previous issue date: 2023-10 | en |
dc.identifier.doi | 10.1007/s00521-023-08932-z | |
dc.identifier.eissn | 1433-3058 | |
dc.identifier.issn | 0941-0643 | |
dc.identifier.uri | https://hdl.handle.net/11693/114873 | |
dc.language.iso | en | |
dc.publisher | Springer | |
dc.relation.isversionof | https://doi.org/10.1007/s00521-023-08932-z | |
dc.rights | CC BY 4.0 Deed (Attribution 4.0 International) | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.source.title | Neural Computing and Applications | |
dc.subject | Activation functions | |
dc.subject | Artificial neural networks | |
dc.subject | Classification | |
dc.subject | Image segmentation | |
dc.title | Deep learning with extended exponential linear unit (DELU) | |
dc.type | Article |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Deep_learning_with_extended_exponential_linear_unit_(DELU).pdf
- Size:
- 2.13 MB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 2.01 KB
- Format:
- Item-specific license agreed upon to submission
- Description: