dc.contributor.author | Karabayır, İbrahim | |
dc.contributor.author | Akbilgiç, Oğuz | |
dc.contributor.author | Taş, Nihat | |
dc.date.accessioned | 2021-12-12T17:01:50Z | |
dc.date.available | 2021-12-12T17:01:50Z | |
dc.date.issued | 2021 | |
dc.identifier.issn | 2162-237X | |
dc.identifier.issn | 2162-2388 | |
dc.identifier.uri | https://doi.org/10.1109/TNNLS.2020.2979121 | |
dc.identifier.uri | https://hdl.handle.net/20.500.11857/3310 | |
dc.description.abstract | Gradient-based algorithms have been widely used in optimizing parameters of deep neural networks' (DNNs) architectures. However, the vanishing gradient remains as one of the common issues in the parameter optimization of such networks. To cope with the vanishing gradient problem, in this article, we propose a novel algorithm, evolved gradient direction optimizer (EVGO), updating the weights of DNNs based on the first-order gradient and a novel hyperplane we introduce. We compare the EVGO algorithm with other gradient-based algorithms, such as gradient descent, RMSProp, Adagrad, momentum, and Adam on the well-known Modified National Institute of Standards and Technology (MNIST) data set for handwritten digit recognition by implementing deep convolutional neural networks. Furthermore, we present empirical evaluations of EVGO on the CIFAR-10 and CIFAR-100 data sets by using the well-known AlexNet and ResNet architectures. Finally, we implement an empirical analysis for EVGO and other algorithms to investigate the behavior of the loss functions. The results show that EVGO outperforms all the algorithms in comparison for all experiments. We conclude that EVGO can be used effectively in the optimization of DNNs, and also, the proposed hyperplane may provide a basis for future optimization algorithms. | en_US |
dc.language.iso | eng | en_US |
dc.publisher | Ieee-Inst Electrical Electronics Engineers Inc | en_US |
dc.relation.ispartof | Ieee Transactions On Neural Networks and Learning Systems | en_US |
dc.identifier.doi | 10.1109/TNNLS.2020.2979121 | |
dc.rights | info:eu-repo/semantics/closedAccess | en_US |
dc.subject | Optimization | en_US |
dc.subject | Training | en_US |
dc.subject | Neural networks | en_US |
dc.subject | Learning systems | en_US |
dc.subject | Handwriting recognition | en_US |
dc.subject | Machine learning algorithms | en_US |
dc.subject | Deep learning | en_US |
dc.subject | CIFAR | en_US |
dc.subject | convolutional neural networks (CNNs) | en_US |
dc.subject | deep learning | en_US |
dc.subject | evolved gradient direction optimizer (EVGO) | en_US |
dc.subject | gradient methods | en_US |
dc.subject | handwritten digit recognition | en_US |
dc.subject | machine learning | en_US |
dc.title | A Novel Learning Algorithm to Optimize Deep Neural Networks: Evolved Gradient Direction Optimizer (EVGO) | en_US |
dc.type | article | |
dc.authorid | akbilgic, oguz/0000-0003-0313-9254 | |
dc.authorid | Karabayir, Ibrahim/0000-0002-7928-176X | |
dc.department | Fakülteler, İktisadi ve İdari Bilimler Fakültesi, Ekonometri Bölümü | |
dc.identifier.volume | 32 | en_US |
dc.identifier.startpage | 685 | en_US |
dc.identifier.issue | 2 | en_US |
dc.identifier.endpage | 694 | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.identifier.wos | WOS:000616310400018 | en_US |
dc.identifier.pmid | PubMed: 32481228 | en_US |
dc.authorwosid | akbilgic, oguz/F-9407-2013 | |
dc.authorwosid | Karabayir, Ibrahim/AAC-3262-2019 | |