Training multi-layer perceptron with enhanced brain storm optimization metaheuristics

  • Nebojsa Bacanin
  • , Khaled Alhazmi
  • , Miodrag Zivkovic
  • , K. Venkatachalam
  • , Timea Bezdan
  • , Jamel Nebhen

Research output: Contribution to journalArticlepeer-review

44 Scopus citations

Abstract

In the domain of artificial neural networks, the learning process represents one of the most challenging tasks. Since the classification accuracy highly depends on the weights and biases, it is crucial to find its optimal or suboptimal values for the problem at hand. However, to a very large search space, it is very difficult to find the proper values of connection weights and biases. Employing traditional optimization algorithms for this issue leads to slow convergence and it is prone to get stuck in the local optima. Most commonly, back-propagation is used for multi-layer-perceptron training and it can lead to vanishing gradient issue. As an alternative approach, stochastic optimization algorithms, such as nature-inspired metaheuristics are more reliable for complex optimization tax, such as finding the proper values of weights and biases for neural network training. In this work, we propose an enhanced brain storm optimization-based algorithm for training neural networks. In the simulations, ten binary classification benchmark datasets with different difficulty levels are used to evaluate the efficiency of the proposed enhanced brain storm optimization algorithm. The results show that the proposed approach is very promising in this domain and it achieved better results than other state-of-the-art approaches on the majority of datasets in terms of classification accuracy and convergence speed, due to the capability of balancing the intensification and diversification and avoiding the local minima. The proposed approach obtained the best accuracy on eight out of ten observed dataset, outperforming all other algorithms by 1-2% on average. When mean accuracy is observed, the proposed algorithm dominated on nine out of ten datasets.

Original languageEnglish
Pages (from-to)4199-4215
Number of pages17
JournalComputers, Materials and Continua
Volume70
Issue number2
DOIs
StatePublished - 2022
Externally publishedYes

Keywords

  • Algorithm hybridization
  • Artificial neural network
  • Brain storm optimization
  • Metaheuristics
  • Optimization

Fingerprint

Dive into the research topics of 'Training multi-layer perceptron with enhanced brain storm optimization metaheuristics'. Together they form a unique fingerprint.

Cite this