TY - JOUR
T1 - HybridHR-Net
T2 - Action Recognition in Video Sequences Using Optimal Deep Learning Fusion Assisted Framework
AU - Akbar, Muhammad Naeem
AU - Khan, Seemab
AU - Farooq, Muhammad Umar
AU - Alhaisoni, Majed
AU - Tariq, Usman
AU - Akram, Muhammad Usman
N1 - Publisher Copyright:
© 2023 Tech Science Press. All rights reserved.
PY - 2023
Y1 - 2023
N2 - The combination of spatiotemporal videos and essential features can improve the performance of human action recognition (HAR); however, the individual type of features usually degrades the performance due to similar actions and complex backgrounds. The deep convolutional neural network has improved performance in recent years for several computer vision applications due to its spatial information. This article proposes a new framework called for video surveillance human action recognition dubbed HybridHR-Net. On a few selected datasets, deep transfer learning is used to pre-trained the EfficientNet-b0 deep learning model. Bayesian optimization is employed for the tuning of hyperparameters of the fine-tuned deep model. Instead of fully connected layer features, we considered the average pooling layer features and performed two feature selection techniques-an improved artificial bee colony and an entropy-based approach. Using a serial nature technique, the features that were selected are combined into a single vector, and then the results are categorized by machine learning classifiers. Five publically accessible datasets have been utilized for the experimental approach and obtained notable accuracy of 97%, 98.7%, 100%, 99.7%, and 96.8%, respectively. Additionally, a comparison of the proposed framework with contemporary methods is done to demonstrate the increase in accuracy.
AB - The combination of spatiotemporal videos and essential features can improve the performance of human action recognition (HAR); however, the individual type of features usually degrades the performance due to similar actions and complex backgrounds. The deep convolutional neural network has improved performance in recent years for several computer vision applications due to its spatial information. This article proposes a new framework called for video surveillance human action recognition dubbed HybridHR-Net. On a few selected datasets, deep transfer learning is used to pre-trained the EfficientNet-b0 deep learning model. Bayesian optimization is employed for the tuning of hyperparameters of the fine-tuned deep model. Instead of fully connected layer features, we considered the average pooling layer features and performed two feature selection techniques-an improved artificial bee colony and an entropy-based approach. Using a serial nature technique, the features that were selected are combined into a single vector, and then the results are categorized by machine learning classifiers. Five publically accessible datasets have been utilized for the experimental approach and obtained notable accuracy of 97%, 98.7%, 100%, 99.7%, and 96.8%, respectively. Additionally, a comparison of the proposed framework with contemporary methods is done to demonstrate the increase in accuracy.
KW - Action recognition
KW - artificial bee colony
KW - deep learning
KW - entropy
KW - feature fusion
KW - transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85174409514&partnerID=8YFLogxK
U2 - 10.32604/cmc.2023.039289
DO - 10.32604/cmc.2023.039289
M3 - Article
AN - SCOPUS:85174409514
SN - 1546-2218
VL - 76
SP - 3275
EP - 3295
JO - Computers, Materials and Continua
JF - Computers, Materials and Continua
IS - 3
ER -