A Spatial And Frequency Based Method For Micro Facial Expressions Recognition Using Color And Depth Images
- https://www.researchgate.net/publication/351334905_A_SPATIAL_AND_FREQUENCY_BASED_METHOD_FOR_MICRO_FACIAL_EXPRESSIONS_RECOGNITION_USING_COLOR_AND_DEPTH_IMAGES_ISSN_2518-8739_Online_JSEIS_httpwwwjseisorgvolumevol6no1html
- Or
- https://figshare.com/articles/journal_contribution/V6N1-2_A_SPATIAL_AND_FREQUENCY_BASED_METHOD_FOR_MICRO_FACIAL_EXPRESSIONS_RECOGNITION_USING_COLOR_AND_DEPTH_IMAGES/14538405?file=27875415
- Or
- https://www.academia.edu/50348966/A_SPATIAL_AND_FREQUENCY_BASED_METHOD_FOR_MICRO_FACIAL_EXPRESSIONS_RECOGNITION_USING_COLOR_AND_DEPTH_IMAGES
- DOI: http://dx.doi.org/10.6084/m9.figshare.14538405
- https://doi.org/10.6084/m9.figshare.14538405.v1
Human face states the inner emotions, thoughts and physical disorders. These emotions are expressed on the face via facial muscles. The estimated time through which a facial expression occurs on the face is between 0.5 to 4 seconds, and a micro expression between 0.1 to 0.5 seconds. Obviously, for the purpose of recording micro expressions, obtaining videos frames between 30 up to 200 frame per second is essential. This research uses Kinect V.2 sensor to get the color and depth data in 30 fps. Depth image stores useful 2.5-Dimentional information from skin wrinkles which is the main key to recognize even slightest micro facial expressions. Experiment starts with splitting color and depth images into facial parts, and after applying preprocessing techniques, features extraction out of both type of data in spatial and frequency domain takes place. Some of the features which are used in this study are Histogram of Oriented Gradient (HOG), Gabor Filter, Speeded Up Robust Features (SURF), Local Phase Quantization (LPQ), Local Binary Pattern (LBP). Non dominated Sorting Genetic Algorithm II (NSGA-II) feature selection algorithm applies on extracted features to have faster learning process and finally selected features are sent to neuro-fuzzy and neural network classifiers. Proposed method is evaluated with the benchmark databases such as, Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME. Also, the proposed method is compared with other similar methods and Convolutional Neural Network (CNN) method on mentioned databases. The results are really satisfactory, and it indicates classification accuracy improvement of proposed method versus other methods.
Key features:
- Input: Color (RGB) and Depth (2.5D) images from Kinect V2.
- Focus on spatial and frequency domain features.
- Benchmark evaluation using six popular datasets: Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME.
-
Feature Extraction:
- Histogram of Oriented Gradients (HOG)
- Gabor Filters
- Speeded Up Robust Features (SURF)
- Local Binary Patterns (LBP)
- Local Phase Quantization (LPQ)
-
Dimensionality Reduction:
- NSGA-II (Non-dominated Sorting Genetic Algorithm II) for evolutionary feature selection.
-
Classifiers:
- Artificial Neural Networks (ANN)
- Neuro-Fuzzy Classifier
-
Comparative Analysis:
- Evaluated against state-of-the-art methods, including CNN and SVM, for both Facial Expression Recognition (FER) and Micro Facial Expressions Recognition (FMER).
- Sensor: Kinect V2 capturing color and depth data at 30 FPS.
- Datasets: Eurecom Kinect Face DB, VAP RGBD-T Face, JAFFE, Face Grabber DB, FEEDB, and CASME.
- Data split into 70% training and 30% testing.
- Facial regions split into mouth, eyes, and nose.
- Applied filters:
- Median low-pass filter for noise reduction.
- Unsharp masking for edge enhancement.
- Histogram Equalization for brightness normalization.
- Canny Edge Detection for feature refinement.
Features extracted from spatial and frequency domains:
- HOG: Captures edge orientations and gradients.
- Gabor Filters: Effective for edge and wrinkle detection.
- SURF: Fast feature detection and robust against rotation.
- LBP: Texture-based feature robust to illumination changes.
- LPQ: Fourier-based features for combating blurring effects.
- NSGA-II: Optimized feature selection for better classification performance and reduced runtime.
- Artificial Neural Networks (ANN):
- 50 hidden layers.
- Conjugate gradient backpropagation for training.
- Neuro-Fuzzy Classifier: