Training of the feed forward artificial neural networks using dragonfly algorithm

dc.contributor.authorGulcu, Saban
dc.date.accessioned2024-02-23T14:02:10Z
dc.date.available2024-02-23T14:02:10Z
dc.date.issued2022
dc.departmentNEÜen_US
dc.description.abstractOne of the most important parts of an artificial neural network (ANN) which affects performance is training algorithms. Training algorithms optimize the weights and biases of the ANN according to the inputs-outputs pattern. Two types of training algorithms are widely used: Gradient methods and meta-heuristic methods. Gradient methods are effective in training the ANN. But they have some disadvantages. The main disadvantage of gradient methods is premature convergence. Secondly, the performance of gradient methods highly depends on the initial parameters and positions. Thirdly, they can easily get stuck in local optima. To overcome these disadvantages, this article presents a new hybrid algorithm (DA-MLP) to train the feed-forward multilayer neural networks (MLP) using the dragonfly algorithm. The dragonfly algorithm optimizes the weights and biases of the MLP. In the experiments, one real-world problem in the civil engineering area and eight classification datasets were used. To verify the success of the DA-MLP algorithm, the results of the DA-MLP algorithm were compared with the results of four algorithms (the BAT-MLP based on the bat optimization algorithm, the SMS-MLP based on the states of matter search optimization algorithm, the PSO-MLP based on the particle swarm optimization algorithm, and the backpropagation (BP) algorithm). The experimental study showed that the DA-MLP algorithm is more efficient than the other algorithms. (c) 2022 Elsevier B.V. All rights reserved.en_US
dc.identifier.doi10.1016/j.asoc.2022.109023
dc.identifier.issn1568-4946
dc.identifier.issn1872-9681
dc.identifier.scopus2-s2.0-85131088632en_US
dc.identifier.scopusqualityQ1en_US
dc.identifier.urihttps://doi.org/10.1016/j.asoc.2022.109023
dc.identifier.urihttps://hdl.handle.net/20.500.12452/11620
dc.identifier.volume124en_US
dc.identifier.wosWOS:000808523400003en_US
dc.identifier.wosqualityQ1en_US
dc.indekslendigikaynakWeb of Scienceen_US
dc.indekslendigikaynakScopusen_US
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.relation.ispartofApplied Soft Computingen_US
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanıen_US
dc.rightsinfo:eu-repo/semantics/closedAccessen_US
dc.subjectArtificial Neural Networksen_US
dc.subjectTraining Of Artificial Neural Networksen_US
dc.subjectDragonfly Algorithmen_US
dc.subjectOptimizationen_US
dc.subjectMultilayer Perceptronen_US
dc.titleTraining of the feed forward artificial neural networks using dragonfly algorithmen_US
dc.typeArticleen_US

Dosyalar