Skip to content

SeyedMuhammadHosseinMousavi/A-SPATIAL-AND-FREQUENCY-BASED-METHOD-FOR-MICRO-FACIAL-EXPRESSIONS-RECOGNITION-USING-COLOR-AND-DEPTH-

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

A Spatial And Frequency Based Method For Micro Facial Expressions Recognition Using Color And Depth Images

Link to the paper:

Please cite:

- MOUSAVI, SEYED MUHAMMAD HOSSEIN. "A SPATIAL AND FREQUENCY BASED METHOD FOR MICRO FACIAL EXPRESSIONS RECOGNITION USING COLOR AND DEPTH IMAGES." Journal of Software Engineering & Intelligent Systems 6.1 (2021): 17.
- Mousavi, Seyed Muhammad Hossein (2021). V6N1-2 A SPATIAL AND FREQUENCY BASED METHOD FOR MICRO FACIAL EXPRESSIONS RECOGNITION USING COLOR AND DEPTH IMAGES. figshare. Journal contribution. https://doi.org/10.6084/m9.figshare.14538405.v1

Micro Facial Expressions Recognition Using Color and Depth Images

This repository contains the implementation of a **spatial and frequency-based method** for **Micro Facial Expressions Recognition (MFER)** using color and depth images captured with Kinect V2. The method leverages advanced feature extraction, dimensionality reduction, and classification techniques to achieve high accuracy in recognizing micro-expressions.
---

Table of Contents


Overview

Human face states the inner emotions, thoughts and physical disorders. These emotions are expressed on the face via facial muscles. The estimated time through which a facial expression occurs on the face is between 0.5 to 4 seconds, and a micro expression between 0.1 to 0.5 seconds. Obviously, for the purpose of recording micro expressions, obtaining videos frames between 30 up to 200 frame per second is essential. This research uses Kinect V.2 sensor to get the color and depth data in 30 fps. Depth image stores useful 2.5-Dimentional information from skin wrinkles which is the main key to recognize even slightest micro facial expressions. Experiment starts with splitting color and depth images into facial parts, and after applying preprocessing techniques, features extraction out of both type of data in spatial and frequency domain takes place. Some of the features which are used in this study are Histogram of Oriented Gradient (HOG), Gabor Filter, Speeded Up Robust Features (SURF), Local Phase Quantization (LPQ), Local Binary Pattern (LBP). Non dominated Sorting Genetic Algorithm II (NSGA-II) feature selection algorithm applies on extracted features to have faster learning process and finally selected features are sent to neuro-fuzzy and neural network classifiers. Proposed method is evaluated with the benchmark databases such as, Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME. Also, the proposed method is compared with other similar methods and Convolutional Neural Network (CNN) method on mentioned databases. The results are really satisfactory, and it indicates classification accuracy improvement of proposed method versus other methods.

Key features:

  • Input: Color (RGB) and Depth (2.5D) images from Kinect V2.
  • Focus on spatial and frequency domain features.
  • Benchmark evaluation using six popular datasets: Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME.

Features

  • Feature Extraction:

    • Histogram of Oriented Gradients (HOG)
    • Gabor Filters
    • Speeded Up Robust Features (SURF)
    • Local Binary Patterns (LBP)
    • Local Phase Quantization (LPQ)
  • Dimensionality Reduction:

    • NSGA-II (Non-dominated Sorting Genetic Algorithm II) for evolutionary feature selection.
  • Classifiers:

    • Artificial Neural Networks (ANN)
    • Neuro-Fuzzy Classifier
  • Comparative Analysis:

    • Evaluated against state-of-the-art methods, including CNN and SVM, for both Facial Expression Recognition (FER) and Micro Facial Expressions Recognition (FMER).

Methodology

Data Acquisition

  • Sensor: Kinect V2 capturing color and depth data at 30 FPS.
  • Datasets: Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME.
  • Data split into 70% training and 30% testing.

Preprocessing

  • Facial regions split into mouth, eyes, and nose.
  • Applied filters:
    • Median low-pass filter for noise reduction.
    • Unsharp masking for edge enhancement.
    • Histogram Equalization for brightness normalization.
    • Canny Edge Detection for feature refinement.

Feature Extraction

Features extracted from spatial and frequency domains:

  • HOG: Captures edge orientations and gradients.
  • Gabor Filters: Effective for edge and wrinkle detection.
  • SURF: Fast feature detection and robust against rotation.
  • LBP: Texture-based feature robust to illumination changes.
  • LPQ: Fourier-based features for combating blurring effects.

Dimensionality Reduction

  • NSGA-II: Optimized feature selection for better classification performance and reduced runtime.

Classification

  • Artificial Neural Networks (ANN):
    • 50 hidden layers.
    • Conjugate gradient backpropagation for training.
  • Neuro-Fuzzy Classifier:
    • Combines neural networks with fuzzy logic.
    • Utilizes fuzzy IF-THEN rules for non-linear approximation. f 2

f3

f5

f7 action units f110

f 13

figure 16

fig 17

f 18

f 19

f 20

res