Skip to main navigation Skip to search Skip to main content

Analysis of Facial Occlusion Challenge in Thermal Images for Human Affective State Recognition

  • Mustafa Al Qudah
  • , Ahmad Mohamed
  • , Syaheerah Lutfi

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

Several studies have been conducted using both visual and thermal facial images to identify human affective states. Despite the advantages of thermal facial images in recognizing spontaneous human affects, few studies have focused on facial occlusion challenges in thermal images, particularly eyeglasses and facial hair occlusion. As a result, three classification models are proposed in this paper to address the problem of thermal occlusion in facial images, with six basic spontaneous emotions being classified. The first proposed model in this paper is based on six main facial regions, including the forehead, tip of the nose, cheeks, mouth, and chin. The second model deconstructs the six main facial regions into multiple subregions to investigate the efficacy of subregions in recognizing the human affective state. The third proposed model in this paper uses selected facial subregions, free of eyeglasses and facial hair (beard, mustaches). Nine statistical features on apex and onset thermal images are implemented. Furthermore, four feature selection techniques with two classification algorithms are proposed for a further investigation. According to the comparative analysis presented in this paper, the results obtained from the three proposed modalities were promising and comparable to those of other studies.

Original languageEnglish
Article number3513
JournalSensors
Volume23
Issue number7
DOIs
StatePublished - Apr 2023
Externally publishedYes

Keywords

  • affects
  • emotion
  • eyeglass
  • occlusion
  • recognition
  • thermal

Fingerprint

Dive into the research topics of 'Analysis of Facial Occlusion Challenge in Thermal Images for Human Affective State Recognition'. Together they form a unique fingerprint.

Cite this