Movatterモバイル変換


[0]ホーム

URL:


Next Article in Journal
Calibrationless Parallel Magnetic Resonance Imaging: A Joint Sparsity Model
Previous Article in Journal
Fabrication of a Microneedle/CNT Hierarchical Micro/Nano Surface Electrochemical Sensor and ItsIn-Vitro Glucose Sensing Characterization
 
 
Search for Articles:
Title / Keyword
Author / Affiliation / Email
Journal
Article Type
 
 
Section
Special Issue
Volume
Issue
Number
Page
 
Logical OperatorOperator
Search Text
Search Type
 
add_circle_outline
remove_circle_outline
 
 
Journals
Sensors
Volume 13
Issue 12
10.3390/s131216682
Font Type:
ArialGeorgiaVerdana
Font Size:
AaAaAa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems

1
UC Lab, Department of Computer Engineering, Kyung Hee University, Yongin-Si 446-701, Korea
2
Division of Information and Computer Engineering, Ajou University, Suwon 443-749, Korea
*
Author to whom correspondence should be addressed.
Sensors2013,13(12), 16682-16713;https://doi.org/10.3390/s131216682
Submission received: 28 October 2013 /Revised: 30 November 2013 /Accepted: 2 December 2013 /Published: 5 December 2013
(This article belongs to the SectionPhysical Sensors)

Abstract

: Over the last decade, human facial expressions recognition (FER) has emerged as an important research area. Several factors make FER a challenging research problem. These include varying light conditions in training and test images; need for automatic and accurate face detection before feature extraction; and high similarity among different expressions that makes it difficult to distinguish these expressions with a high accuracy. This work implements a hierarchical linear discriminant analysis-based facial expressions recognition (HL-FER) system to tackle these problems. Unlike the previous systems, the HL-FER uses a pre-processing step to eliminate light effects, incorporates a new automatic face detection scheme, employs methods to extract both global and local features, and utilizes a HL-FER to overcome the problem of high similarity among different expressions. Unlike most of the previous works that were evaluated using a single dataset, the performance of the HL-FER is assessed using three publicly available datasets under three different experimental settings: n-fold cross validation based on subjects for each dataset separately; n-fold cross validation rule based on datasets; and, finally, a last set of experiments to assess the effectiveness of each module of the HL-FER separately. Weighted average recognition accuracy of 98.7% across three different datasets, using three classifiers, indicates the success of employing the HL-FER for human FER.

    1. Introduction

    Over the last decade automatic FER has become an important research area for many applications, such as more engaging human–computer interfaces; image retrieval; human emotion analysis [1]; neuroscience, psychology, and cognitive sciences [2]; access control and surveillance [3]; and communication, personality, and child development [4].

    Human FER systems can be classified into two categories; pose-based expression recognition systems [57] and spontaneous expression recognition systems [810]. Pose-based expressions are the artificial expressions produced by people when they are asked to do so [2]. Similarly, spontaneous expressions are those that people give out spontaneously, and they can be observed on a day-to-day basis, such as during conservations or while watching movies [2]. The focus of this article is pose-based FER systems.

    As shown inFigure 1, a typical FER system consists of three basic modules: pre-processing, feature extraction, and recognition. The pre-processing module performs two tasks. Firstly, it diminishes illumination and other light effects to increase the recognition accuracy, using techniques like morphological filters, homomorphic filters, or median filters. Most previously FER systems [1115] have exhibited low accuracy due to the lack of this filtering element in their architecture. One of the most well-known methods that have been utilized to diminish such illumination effects is histogram equalization (HE). However, in a gray level image HE assigns the highest gray level value to each pixel [16], as a result the resulting image produced by HE contains gapsi.e., “empty-bins” between very full histogram bins [17]. Due to this limitation, HE causes unwanted artifacts and a washed-out look, so it is not recommended [18,19].

    Therefore, two improved versions of HE were proposed in order to solve the limitations of HE. The first technique is Local Histogram Equalization (LHE) that uses a sliding window technique in order to compute the local histogram for each pixel and the gray level for center pixel is changed accordingly [17]. LHE, though better than HE, causes over-enhancement, and sometimes it produces checkerboards of the enhanced image [20] and it requires more time than other methods [21]. The second technique is Global Histogram Equalization (GHE), which uses the histogram information of the whole image; therefore, we employed GHE instead of using HE and LHE in preprocessing module. To the best of our knowledge, it is the first time that GHE is being used for facial expression recognition.

    Naturally, before recognizing facial expressions a face is must be located in the image/video frame. Therefore, the second task of the preprocessing module is to detect faces in a given image, as it is the face that contains most of the expression-related information. For this, many well-known methods, including the appearance-based methods [22,23], have been proposed. The appearance-based methods have shown excellent performance in a static environment; however, their performance degrades when used in dynamic situations [24]. Other commonly used face detection methods include neural network-based face detection [25], and digital curves-based face detection [26]. These methods utilize eye window (eye regions) in order to locate a face. However, eyes regions are highly sensitive to the hairstyle of the person, which may cause misclassification [27]. Moreover, in their methods, it is very hard to know the orientation of the detected face [27]. Many existing works, such as [2830], have performed facial expression recognition without face detection, making these systems heuristic in nature. To solve this problem, some face detection and extraction techniques were proposed [3133]. In [31], the authors detected the human face using the position of the eyes and then cropped the faces accordingly. However, this method failed to detect the face structure in very sensitive environmental conditions. In contrast, in [33] features for face detection were selected manually.

    PCA has been exploited by [34,35] to extract the features from eyes and the lips, which were then employed by a framework of radial basis function network (RBFN) to classify the expressions based on the extracted features. PCA extracts only the global features; therefore, another well-known higher statistical order technique named ICA has also been widely employed in FER systems to extract the local features. The authors of [36] employed supervised ICA for feature extraction in their FER system and showed improved recognition accuracies. As ICA is an unsupervised technique; therefore, in their work, they modified the classical ICA and made it a supervised ICA (sICA) by including a prior knowledge that was reassembled from the training data using a Maximum Posteriori (MAP) scheme. Other important feature extraction technique used for the sake of FER in the past include: non-negative matrix factorization (NMF) and Local Non-negative Matrix Factorization (LNNF) [37], Higher order Local Auto-Correlation (HLAC) [38], Local Binary Pattern (LBP) [13], Gabor wavelet [39], and BoostedICA [40].

    Regarding the feature extraction module, several methods have been explored. Some of the most well-known features used for FER include texture-based features [41], geometry-based features [42], holistic features such as nearest features using line-based subspace analysis [43], Eigenfaces and Eigenvector [4446], Fisherfaces [47], global features [48], and appearance based features [49,50]. In this article, we utilized the holistic feature extraction methods,i.e., Principal Component Analysis (PCA) and Independent Component Analysis (ICA) in order to extract the prominent features from the expression frames.

    As for the recognition module, several classifiers have been investigated. The authors of [51] employed artificial neural networks (ANNs) to recognize different types of facial expressions and achieved an accuracy of 70%. However, an ANN is a black box and has incomplete capability to explicitly categorize possible fundamental relationships [52]. Moreover, the FER systems proposed in [8,5357] used support vector machines (SVMs). However, in SVMs, the probability is calculated using indirect techniques; in other words, there is no direct estimation of the probability, these are calculated by employing five-fold cross validation due to which SVM suffers from the lack of classification [58]. Furthermore, SVMs simply disregard the temporal addictions among video frames, thus each frame is expected to be statistically independent from the rest. Similarly, in [49,59,60], Gaussian mixture models (GMMs) were employed to recognize different types of facial expressions. As stated earlier, the features could be very sensitive to noise; therefore, fast variations in the facial frames cannot be modeled by GMMs and produces problems for sensitive detection [61]. Hidden Markov Models (HMMs) are mostly used to handle sequential data when frame-level features are used. In such cases, other vector-based classifiers like GMMs, ANNs, and SVMs, have difficulty in learning the dependencies in a given sequence of frames. Due to this capability, some well-known FER systems, including [6264], utilized HMM as a classifier. In conclusion, a large number of feature extraction techniques and classifiers have been employed for video-based FER systems. Among them, PCA and ICA have been the most widely used feature extraction techniques, and HMMs have been the most commonly used classifier.

    A recent work by Ziaet al. [64] proposed a complete approach for FER systems that provided high classification accuracy for the Cohn-Kanade database of facial expressions. In their work, they employed PCA and ICA for feature extraction. Once extracted, features were subject to Linear Discriminant Analysis (LDA) to find the most relevant features. The result after applying LDA was fed to an HMM. The recognition rate of their technique was 93.23% when tested on Cohn-Kanade dataset. However, their technique failed in exhibiting the same accuracy when tested by us on other datasets, such as Japanese Female Facial Expression (JAFFE) database (83%), and AT&T database (72%) of facial expressions. Low accuracy in these new experiments could be attributed to the following two reasons. Firstly, in their work, they did not use a pre-processing step to diminish the lighting and illumination effects. Furthermore, in some of the databases such as the AT&T database of facial expressions, the subjects have worn glasses that make it difficult to extract useful features from some parts of the face, such as the eyes.

    Secondly, most of the expressions share high similarity, and thus their features overlap significantly in the feature space, as shown inFigure 2. Ziaet al. [64] applied LDA to the extracted feature space to improve the class separation among different classes with the assumption that the variance is distributed uniformly among all the classes. However, this is not the case. For example, expressions like happiness and sadness are very similar to each other but can easily be distinguished from anger and fear (another pair with high similarity). Accordingly, this work implements a HL-FER that is capable of performing accurate facial expression recognition across multiple datasets. Previously, such model has been used in [65] that was dual-layer SVM ensemble classification. The motivation behind their study was to determine how the contraction of muscles changes the appearance of the face by extracting the local features from the three parts of the face such as mouth, nose, and eyes. However, the performance of the dual-layer SVM classification cannot match that of binary classification as SVMs use approximation algorithms in order to decrease the computation complexity but these have the effect of degrading classification performance [66].

    In our HL-FER, firstly, images were passed through a pre-processing module to diminish the illumination effects, and to extract the human face automatically. Secondly, PCA and ICA were used for feature extraction. Finally, a hierarchical classifier was used, where the expression category was recognized at the first level, followed by the actual expression recognition at the second level. The HL-FER has been validated using three different experiments. The results of these experiments show that the two-level recognition scheme, along with the proposed pre-processing module for noise reduction and accurate face detection, solved the aforementioned problems of the existing FER systems; and therefore, succeeded in providing high recognition accuracy across multiple datasets.

    Above, we discussed some related work in this field. The rest of the paper is organized as follows. Section 2 provides an overview of our HL-FER. Section 3 discusses the experimental setup along with the experimental results with some discussion on the results and talks about the factors that could degrade systems performance if tested in real-life scenarios. Section 4 provides the analysis and comparison of the recognition accuracy of this work with those of the some of the existing FER systems. Finally, the paper is concluded with some future directions in Section 5.

    2. Materials and Methods

    The architecture of our HL-FER is shown inFigure 3.

    2.1. Pre-Processing

    As mentioned earlier, in most of the datasets, the face images have various resolution and backgrounds, and were taken under varying light conditions; therefore, pre-processing module is necessary to improve the quality of images. At this stage, background information, illumination noise, and unnecessary details are removed to enable fast and easy processing. After employing this stage, we can obtain sequences of images which have normalized intensity, size and shape. Several techniques exist in literature to diminish such illumination effects, such methods are local histogram equalization (LHE) and global histogram equalization (GHE). However, LHE causes over-enhancement, and sometimes it produces checkerboards of the enhanced image [38]. Therefore, we employed a GHE for diminishing illumination effects. To the best of our knowledge, this is the first time GHE has been exploited for facial expression recognition. For more detail on GHE, please refer to [39].

    In the proposed face detection algorithm, two improved key face detection methods were used simultaneously: gray-level and skin-tone-based. To attain the best performance, the similarity angle measurement (SAM) method was used. SAM utilizes two energy functions,F(C) andB(C), to minimize the dissimilarities within the face and maximize the distance between the face and the background, respectively. The overall energy function can be defined as:

    E(C)=βF(C)+(1β)B(C)
    Whereβ ∈ [0−1], and
    F(C)=inside(C)|Icin|2dx+outside(C)|Icout|2dx
    wherecin andcout are the average intensities inside and outside the variable boundaryC, andI is the facial image. Furthermore:
    B(C)=zf1(x)f2(x)dx
    where:
    f1(x)=Kσ(x)*[H(ϕ(x))I]Kσ(x)*[H(ϕ(x))]
    and:
    f2(x)=Kσ(x)*[H(ϕ(x))I]Kσ(x)*[1H(ϕ(x))]
    wheref1(x) andf2(x) are the local fitting functions, which depend on the facial geometry functionϕ(x), and need to be updated in each iteration,x is the corresponding region, andH(•) andKσ(•) as the Heaviside and Dirac functions, respectively.

    In summary, the intuition behind the proposed energy functional for SAM is that we seek for a boundary which partitions the facial image into regions such that the differences within each region are minimized (i.e., theF(C) term) and the distance between the two regions (face and background) is maximized (i.e., theB(C) term). The facial geometry functionϕ(x) implementation for the energy functional inEquations (4) and(5) is carried out in order to define the boundary between the two regions (face and background) and can be derived as:

    ϕ(x)=ϕt=|ϕ|{γKσ+η+β[(Icin)2+(Icout)2](1β)[B(C)2(1Ain1Aout)+12zKσ(z1)(1Aoutf1(x)f2(x)1Ainf1(x)f2(x))dz]}
    Where |∇ϕ| is the gradient for the facial geometry functionϕ(x), andAin andAout are respectively the areas inside and outside of the boundaryC,γ,β andη are the three parameters, such thatγ helps to detect objects of various sizes, including small points caused by noise;β weights the constraints of within-face homogeneity and between-face dissimilarity for which the value ofβ should be small;η speeds up the function evolution but may make it to pass through weak edges. Therefore, we usedη = 0 in all experiments for fair comparison. Once the boundary between face and background is found, then in the next step, the skin-tone and gray-level methods are collectively applied in order to accurately detect and extract the face from the facial frame. For further details on gray-level and skin-tone methods please refer to [27,67], respectively.

    In summary, the proposed face detection system is based on multiple features from a face image/frame. When a rough face image is presented to the system, an improved gray-level and skin-tone model is adopted to locate the face region. Very often, hair is included in the detected head contour. The second step is to find the precise face geometry using a facial geometry operation. To locate the exact face region, three features were used: the face intensity (because the intensity of the eye region is relatively low [27]); the direction of the line connecting the center of the eyes, which is determined on the face edge image; and the response of convolving the proposed facial geometry variance filter with the face image. We have developed a facial geometry filter for extracting potential face windows, based on similarity measurements of facial geometry signatures. This process generates a list of possible face signatures, since each face feature has unique identification, or signatures. These signatures can be compared with face signatures that are stored in the database to determine positive matches. The stored face signatures were in true (original) form and were not distorted. However, in some cases, the detected face boundary usually might consist of hair [68]. Hairstyles are different from person to person [69]. Therefore, in the next step another technique,i.e., face skin region has been employed in order to locate the skin region from the detected face region and to get rid of the problem of different hairstyles.

    The facial features such as nose, mouth, and eyes possess relatively lower gray levels under normal illumination [27]. We can always plot the intensity histogram of the face image, because skin color has a relatively high gray intensity, while other facial components have relatively low intensities, and by this way, it is easy to find a threshold value for face detection under normal illumination. This threshold value, which is obtained by the intensity histogram, is used as the criterion for successful face detection. In most images, the number of possible cases for the second iteration of the loop (seeFigure 4) was less than two. For each possible case, the signature similarity measurement (SSM) function was adopted for face detection, tracking, and verification. If the face was detected in the face frame, the detection process completed; if not, next possible case was tested. Details of each block are presented inFigure 4.

    2.2. Feature Extraction

    As described earlier, numerous techniques have been developed and validated for the purpose of feature extraction for FER systems. Among these, PCA and ICA are the mostly commonly used methods, and their performance has already been validated in [64]. Therefore, we decided to use PCA and ICA for feature extraction to extract both the global and local features respectively.

    2.3. Recognizing the Expression Category

    This work is based on the theory that different expressions can be grouped into three categories based on the parts of the face that contribute most toward the expression [7072]. This classification is shownTable 1.

    Lips-based expressions are those in which the lips make up the majority of the expression. In lips-eyes-based expressions, both lips and eyes contribute in the expression. In lips-eyes-forehead expressions, lips, eyes, and eyebrows or forehead have equal roles. In the HL-FER, an expression is classified into one of these three categories at the first level. At the second level, classifier (trained for the recognized category) is employed to give a label to this expression within this category.

    At the first level, LDA was firstly applied to the extracted features from all the classes and an HMM was trained to recognize the three expression categories: lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions. The LDA-features for these three categories are shown inFigure 5. A clear separation could be seen among the categories, and this is why the HL-FER achieved 100% recognition accuracy at the first level.

    2.4. Recognizing the Expressions

    As mentioned earlier, once the category of the given expression has been determined, the label for the expression within the recognized category is recognized at the second level. For this purpose, LDA was applied separately to the feature space of each category and the result was used to train three HMMs, one HMM per category. Collectively, the overall results for all the expression classes are shown inFigure 6.

    These feature plots indicate that applying LDA to the features of three categories separately provided a much better separation as compared to single-LDA via single-HMM approach (seeFigure 7). The single-LDA via single-HMM approach means that instead of applying LDA separately to each expression category and using separate HMMs for these categories, LDA is applied only once, to the features of all the classes, and only one HMM is used for classification. ForFigures 5,6 and7, we used Cohn-Kanade dataset, forFigures 8,9 and10, we employed JAFFE dataset, and forFigures 11,12 and13, we used AT&T dataset. Each dataset consisted of six basic universal facial expressions and each expression has twelve facial expression frames.

    We have sequential data at both levels; therefore at both levels HHMs have been employed, because HMMs have their own advantage of handling sequential data when frame-level features are used. In such cases, other vector-based classifiers like GMMs, ANNs, and SVMs, have difficulty in learning the dependencies in a given sequence of frames. The following formula that has been utilized in order to model HMM (λ):

    λ=(O,Q,π)
    whereO is the sequence of observations e.g.,O1,O2,…,OT and each state is denoted byQ such asQ =q1,q2,…,qN, whereN is the number of the states in the model, andπ is the initial state probabilities. The parameters that used to model HMM for all experiments were 44, 4, and 4 respectively.

    3. Experimental Results

    3.1. Setup

    In order to assess the HL-FER, six universal expressions like: happiness, sadness, surprise, disgust, anger, and fear were used from three publicly available standard datasets, namely the Cohn-Kanade [73], JAFFE [74] and AT&T [75] datasets. These datasets display the frontal view of the face, and each expression is composed of several sequences of expression frames. During each experiment, we reduced the size of each input image (expression frame) to 60 × 60, where the images were fi wh converted to a zero-mean vector of size 1 × 3,600 for feature extraction. All the experiments were performed in Matlab using a dual-core Pentium processor (2.5 GHz) with a RAM capacity of 3 GB. Some information on the three datasets is as follows:

    3.1.1. Cohn-Kanade Dataset

    In this facial expressions dataset, there were 100 subjects (university students) performed basic six expressions. The age range of the subjects were from 18 to 30 years and most of them were female. We employed those expression for which the camera was fixed in front of the subjects. By the given instructions, the subjects performed a series of 23 facial displays. Six expressions were based on descriptions of prototypic emotions such as happy, anger, sad, surprise, disgust, and fear. In order to utilize these six expressions from this dataset, we employed total 450 image sequences from 100 subjects, and each of them was considered as one of the six basic universal expressions. The size of each facial frame was 640 × 480 or 640 × 490 pixel with 8-bit precision for grayscale values. For recognition purpose, twelve expression frames were taken from each expression sequence, which resulted in a total of 5,400 expression images.

    3.1.2. JAFFE Dataset

    We also employed Japanese Female Facial Expressions (JAFFE) dataset in order to assess the performance of the HL-FER. The expressions in the dataset were posed by 10 different (Japanese female) subjects. Each image has been rated on six expression adjectives by 60 Japanese subjects. Most of the expression frames were taken from the frontal view of the camera with tied hair in order to expose all the sensitive regions of the face. In the whole dataset, there were total 213 facial frames, which consists of seven expressions including neutral. Therefore, we selected 205 expression frames for six facial expressions performed by ten different Japanese female as subjects. The size of each facial frame was 256 × 256 pixels.

    3.1.3. AT&T Dataset

    Additionally, we also employed AT&T dataset of facial expressions to evaluate the performance of the HL-FER. There are 10 facial frames in each expression performed by 40 distinct subjects. The frames were taken at different illuminations of light against a dark homogenous background with the subjects, and were in grey scale having size of 92 × 112 pixels. This dataset consists of open/close eyes, smiling and not smiling expressions. Among these expressions, few expressions that showed the basic six expressions; therefore, we have manually chosen those suitable facial frames for our experiments. The total numbers of selected facial frames were 240.

    Moreover, in order to show the efficacy of the HL-FER, we performed some more experiments on Yale database of facial expressions [76]. The results are shown in theAppendix inFigures A1,A2,A3, and A4 and inTable A5 respectively.

    3.2. Results and Discussion

    3.2.1. Experimental Face Detection Method Results

    Some samples results for the proposed preprocessing method (GHE) along with the proposed face detection are shown inFigure 14.Figure 14b indicates the success of the proposed GHE in diminishing the lighting effects from both a bright and dark image (shown inFigure 14a) in order to enhance the facial features. After this, in the next step, the proposed face detection method found the facial faces in the frames (shown inFigure 14c) and cropped it accordingly, as shown inFigure 14d.

    The detection rates of the proposed face detection algorithm for the three datasets are summarized inTable 2. It can be seen fromTable 2 that the proposed face detection algorithm achieved high recognition rate on the three different standard datasets of facial expressions.

    3.2.2. Experimental Results of HL-FER Based on Subjects

    The experiments for the HL-FER were performed in this order. In the fin t experiment, the HL-FER was validated on three different datasets. Each dataset possessed different facial features, such as some of the subjects have worn glasses in AT&T dataset while the subjects of the Cohn-Kanade and JAFFE datasets are free of glasses. The remaining facial features of AT&T and Cohn-Kanade datasets are quite similar with each other. On the other hand, the facial features, such as the eyes of the subjects in the JAFFE dataset are totally different from the eyes of the subjects of AT&T and Cohn-Kanade datasets [77,78].The HL-FER was evaluated for each dataset separately that means for each dataset,n-fold cross-validation rule (based on subjects) was applied. It means that out ofn subjects, data from a single subject was retained as the validation data for testing the HL-FER, whereas the data for the remainingn − 1 subjects were used as the training data. This process was repeatedn times, with data from each subject used exactly once as the validation data. The value ofn varied according to the dataset used. The detailed results of this experiment for the three datasets are shown inTables 3,4 and5, respectively.

    Despite of all these differences, the HL-FER consistently achieved a high recognition rate when applied on these datasets separately,i.e., 98.87% on Cohn-Kanade, 98.80% on JAFFE, and 98.50% on the AT&T dataset. This means that, unlike Ziaet al. [64], the HL-FER is robusti.e., the system not only achieves high recognition rate on one dataset but shows the same performance on other datasets as well. The reference images of the three datasets are shown inFigure 15.

    3.2.3. Experimental Results of HL-FER Based on Datasets

    In the second experiment the HL-FER cross-dataset validation was performed. This means that from the three datasets, data from the two datasets were retained as the validation data for testing the system, and the data from the remaining dataset was used as the training data. This process was repeated three times, with data from each dataset used exactly once as the training data.

    The experimental results of the HL-FER at the first level classification for the category recognition on Cohn-Kanade dataset is shown inTable 6, while on the JAFFE dataset, the results are indicated inTable 7, and for the AT&T dataset, the results are represented inTable 8. Similarly, the experimental results of the HL-FER at the second level are summarized inTables 9,10, and11 respectively.

    3.2.4. Experimental Results of HL-FER under the Absence of Each Module

    Thirdly, a set of experiments was performed to assess the effectiveness of each module of the HL-FER (pre-processing, face detection and hierarchical recognition) separately. This experiment was repeated three times and the recognition performance was analyzed under three different settings: Firstly, the experiment was repeated without the pre-processing step. Secondly, the experiment was performed without including the face detection module. In this case the accuracy for the HL-FER is same as indicated inTables 3,4 and5,i.e., when the proposed face detection method fails to detect the face in the facial frame, then there is no effect on the accuracy of the HL-FER, but the HL-FER processes the whole frame instead processing the region of interest (face); however, it will take a bit more time to recognize the expression by considering the whole frame. And, lastly, a single LDA and HMM were used to recognize all the expressions instead of using the HL-FER. The results for the three settings on the Cohn-Kanade dataset are shown inTables 12,3 and13, on the JAFFE dataset are presented inTables 14,4, and15, and on the AT&T dataset are displayed inTables 16,5, and17, respectively.

    It can be seen fromTables 12,13,14,15,16 and17 that both the preprocessing and hierarchical recognition modules of the HL-FER are important. As indicated inTables 13,15, and17, the hierarchical recognition is mainly responsible for the high recognition accuracy of the HL-FER. When we removed the hierarchical recognition module, the recognition rate decreased significantly. These results support the theory that the problem of high similarity among the features of different expressions is a local problem. In other words, the features exist in the form of groups in the overall feature space. The expressions within one group are very similar, whereas they are easily distinguishable from those in the other groups; therefore, to overcome this problem in an effective manner, these groups (or expression categories) should be separated first and then techniques like LDA should be applied to each category separately.

    In the next experiment, the accuracy of the HL-FER when tested across different datasets decreased significantly. We believe that this decrease in accuracy is due to training the HL-FER using the JAFFE dataset and then testing the HL-FER on the AT&T or Cohn-Kanade datasets, andvice versa. As explained earlier, the facial structures, especially eyes, of the subjects in the Japanese dataset are very different than those of the AT&T and Cohn-Kanade datasets [79], which acts as noise and thus degrades the HL-FER performance. To test this theory, another experiment was performed where the HL-FER was first trained using the Cohn-Kanade and then tested on the AT&T dataset, and then the same experiment was repeated while switching the datasets. The results of these experiments are shown inTables 18 and19.

    It can be seen from theTables 18 and19 that the HL-FER achieved good results and proved the early stated theory of lower recognition accuracy due to the difference in the facial features of the subjects in the JAFFE dataset and the other two datasets.

    4. Comparison and Analysis of HL-FER

    The performance of the HL-FER was compared against nine conventional methods [3440,64,79], for all the three datasets,i.e., the Cohn-Kanade, JAFFE, and AT&T datasets of facial expressions. All of these methods were implemented by us using the instructions provided in their respective papers. For each dataset,n-fold cross-validation rule (based on subjects) was applied. In other words, out ofn subjects, data from a single subject was retained as the validation data for testing the HL-FER, whereas the data for the remainingn − 1 subjects were used as the training data. This process was repeatedn times, where the value ofn varied according to the dataset used. The average recognition results (for the three datasets) for each of these conventional methods are listed inTable 20. It can be seen that the HL-FER outperformed the existing facial expression systems.

    Lastly, to analyze the computational cost of the HL-FER, its average recognition time for the three datasets was compared to that of [64], as [64] achieved the highest recognition accuracy in the above experiment among the nine conventional methods. The average computational time taken by [64] for the Cohn-Kanade, JAFFE, and AT&T datasets was 662, 375, and 590 ms, respectively. The datasets had 5,400, 205, and 240 images, respectively. On the other hand, the HL-FER took 1,655 ms for the Cohn-Kanade dataset, 639 ms for the JAFFE dataset, and 1,180 ms for the AT&T dataset. These results show that though the HL-FER showed significant improvement over conventional methods in terms of recognition accuracy, it achieved that at the expense of a higher computational cost. In the HL-FER, we used two level classifications and in each level, we utilized HMM for classification; therefore, the HL-FER took a bit more time than of the existing method [64].

    All these experiments were performed in the laboratory (offline validation) using three standard datasets. Though the system performed accurately and robustly in all the experiments, the effects on system performance once implemented in real time are yet to be investigated. There exist several elements that could degrade the systems performance, such as the clutter and varying face angles. Clutter means that there could be some unnecessary objects in the images along with the test subject. Solving such problems would require a real-time robust segmentation technique. Moreover, the images used in this work were taken only from the frontal view; however, in real life the angles of the camera might varyi.e., side views can also happen. Therefore, further study is needed to tackle these issues and maintain the same high recognition rate in real-life environment.

    Eventually, we envision that the HL-FER will be employed in smartphones. Even though our system showed high accuracy, it employs two-level-recognition with HMMs used at each level. This might become a complexity issue, especially when used in smartphones. One solution could be to use a lightweight classifier such as k-nearest neighbor (k-NN) at the first level; however k-NN has its own limitations such as it is very sensitive to the presence of inappropriate parameters and sensitive to noise as well. Therefore, it can have poor performance in a real-time environment if the training set is large. Therefore, further study is needed to find ways to maintain the high accuracy of the HL-FER while improving its efficiency at the same time to be used in smartphones.

    5. Conclusions

    Over the last decade, automatic human FER has become an important research area for many applications. Several factors make FER a challenging research problem, such as varying light conditions in training and test images, the need for automatic and accurate detection of faces before feature extraction, and the high similarity among different expressions resulting in overlaps among feature values of different classes in the feature space. That is why, though several FER systems have been proposed that showed promising results for a certain dataset, their performance was significantly reduced when tested with different datasets. In this paper, we proposed and validated the accuracy of a HL-FER. Unlike the previous systems, the HL-FER involved a pre-processing step based on global histogram equalization (GHE), which helped in provide high accuracy by eliminating the light effects. We also proposed a face detection technique that utilizes both gray levels and skin-tones to automatically detect faces with high accuracy. Further, we employed both PCA and ICA to extract both the global and the local features. Finally, we used a HL-FER to overcome the problem of high similarity among different expressions. Expressions were divided into three categories based on different parts of the face. At the first level, LDA was used with an HMM to recognize the expression category. At the second level, the label for an expression within the recognized category is recognized using a separate set of LDA and HMM, trained just for that category. In order to evaluate the performance of the HL-FER detailed experiments were conducted on three publicly available datasets in three different experimental settings. The HL-FER achieved a weighted average recognition accuracy of 98.7% using three HMMs, one for per category expression across three different datasets (the Cohn-Kanade dataset has 5,400 images, the JAFFE dataset has 205 images, and the AT&T dataset has 240 images), illustrating the successful use of the HL-FER for automatic FER.

    Acknowledgments

    This work was supported by a post-doctoral fellowship grant from the Kyung Hee University Korea in 2009 [KHU-20090439].

    Appendix

    This sections provides an overview of the results of different other experiments performed during the course of the development of this work.

    Table
    Table A1. Confusion matrix of PCA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    Table A1. Confusion matrix of PCA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    CKJAATCKJAATCKJAATCKJAATCKJAATCKJAAT
    Happiness63.55550.55.59.915.56.57.36.35.38.59.19.2119.7108.38.9
    Sadness6.3132155.250.34510.56.98.39.4129.95.98.97.712.78.98.1
    Anger4.56.78.912.59.810.953.94851.811.910.79.56.39.88.910.91510
    Disgust4.56.89.48.37.810.89.58.91265.855.8407.714214.26.76.8
    Surprise8.98.586.95.67.98.68.78.99.81011.660.16055.95.77.27.7
    Fear9.66.77.811.810.79.110.513159.78.48.48.58.21249.95347.7
    Average56.3350.1651.2353.8558.6650.2
    Table
    Table A2. Confusion matrix of PCA + LDA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    Table A2. Confusion matrix of PCA + LDA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    CKJAATCKJAATCKJAATCKJAATCKJAATCKJAAT
    Happiness68.170.969.17.989.15.65.46.27.95.95.54.73.65.85.86.24.3
    Sadness6.88.81072.369.565.57.264.85.84.87.9455.83.95.96
    Anger4.93.23.37.94.85.170.974.9755.85.53.45.64.65.24.978
    Disgust3.96.74.38.57.97.57.87.86.568.56166.75.6108.95.76.66.1
    Surprise6.13.35.74.34.23.25.23.46.86.18.1973.678.9714.72.14.3
    Fear5.83.93.66.65.76.16.97.98.15.85.74.45.95.84.8697173
    Average69.3669.1074.2465.4074.5071.00
    Table
    Table A3. Confusion matrix of ICA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    Table A3. Confusion matrix of ICA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    CKJAATCKJAATCKJAATCKJAATCKJAATCKJAAT
    Happiness69.265.7636.29.5105.44.66.85.16.97.96.05.75.98.17.66.4
    Sadness2.89.6971.36971.78.63.75.15.84.83.87.76.96.13.864.3
    Anger4.73.64.38.74.95.673.977.473.95.93.84.604.33.76.867.9
    Disgust4.943.78.764.97.96.25.565.871.569.46.87.79.95.94.66.6
    Surprise3.253.64.75.93.45.36.46.13.98.9977.169755.84.82.9
    Fear5.82.648.73.84.17.987.66.54.55.36.75.36.264.475.872.8
    Average65.9670.6675.0668.9073.7071.00
    Table
    Table A4. Confusion matrix of ICA + LDA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    Table A4. Confusion matrix of ICA + LDA and HMM using Cohn-Kanade (CK), JAFFE (JA), and AT&T (AT) databases of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    CKJAATCKJAATCKJAATCKJAATCKJAATCKJAAT
    Happiness73.075.770.85.87.2115.84.706.906.205.27.28.57.24.8
    Sadness4.26.97.175.173756.75.24.26.25.962.54.355.34.72.7
    Anger34.65.28.95.23.369.971.2766.74.84.86.785.44.86.25.3
    Disgust06.75.56.606.49.28.47.275.869.972.88.498.1060
    Surprise6.533.24.83.22.63.53.83.1067.379.38081.45.942.4
    Fear4.5007.8035.68.9104.96.6707.8377.276.777
    Average73.1674.3672.3672.8384.476.96
    Sensors 13 16682f16 1024
    Figure A1. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions on Yale B dataset for six different types of facial expressions. It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Figure A1. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions on Yale B dataset for six different types of facial expressions. It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Sensors 13 16682f16 1024
    Sensors 13 16682f17 1024
    Figure A2. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category on Yale B dataset for six different types of facial expressions. It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 7.
    Figure A2. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category on Yale B dataset for six different types of facial expressions. It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 7.
    Sensors 13 16682f17 1024
    Sensors 13 16682f18 1024
    Figure A3. 3D-feature plot of single-LDA via single-HMM on Yale B dataset for six different types of facial expressions. It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 6).
    Figure A3. 3D-feature plot of single-LDA via single-HMM on Yale B dataset for six different types of facial expressions. It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 6).
    Sensors 13 16682f18 1024
    Table
    Table A5. Confusion matrix for the HL-FER using Yale B dataset of facial expressions (Unit: %).
    Table A5. Confusion matrix for the HL-FER using Yale B dataset of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9910000
    Sadness3970000
    Anger0098002
    Disgust0009910
    Surprise0003970
    Fear0020098
    Average98.00

    Conflicts of Interest

    The authors declare no conflict of interest.

    References

    1. Fasel, B.; Luettin, J. Automatic facial expression analysis: A survey.Patt. Recognit.2003,36, 259–275. [Google Scholar]
    2. Bettadapura, V.Face Expression Recognition and Analysis: The State of the Art; Tech Report arXiv:1203.6722; April; 2012; pp. 1–27. [Google Scholar]
    3. Wu, X.; Zhao, J. Curvelet Feature Extraction for Face Recognition and Facial Expression Recognition. Proceedings of IEEE 6th International Conference on Natural Computation, Yantai, China, 10–12 August 2010; Volume 3, pp. 1212–1216.
    4. Bartlett, M.; Hager, J.; Ekman, P.; Sejnowski, T. Measuring facial expressions by computer image analysis.Psychophysiology1999,36, 253–263. [Google Scholar]
    5. Moore, S.; Bowden, R.; Guildford, U. The Effects of Pose on Facial Expression Recognition. Proceedings of BMVC, London, UK, 7‱10 September 2009; pp. 1–11.
    6. Gross, R.; Baker, S.; Matthews, I.; Kanade, T. Face Recognition across Pose and Illumination. InHandbook of Face Recognition; Springer: New York, NY, USA, 2005; pp. 193–216. [Google Scholar]
    7. Zhu, Z.; Ji, Q. Robust Real-Time Face Pose and Facial Expression Recovery. Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, USA, 17–22 June 2006; Volume 1, pp. 681–688.
    8. Bartlett, M.; Littlewort, G.; Frank, M.; Lainscsek, C.; Fasel, I.; Movellan, J. Recognizing Facial Expression: Machine Learning and Application to Spontaneous Behavior. Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 568–573.
    9. Bartlett, M.S.; Littlewort, G.; Braathen, P.; Sejnowski, T.J.; Movellan, J.R. A Prototype for Automatic Recognition of Spontaneous Facial Actions. InAdvances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2003; Volume 15, pp. 1271–1278. [Google Scholar]
    10. Bartlett, M.S.; Littlewort, G.C.; Frank, M.G.; Lainscsek, C.; Fasel, I.R.; Movellan, J.R. Automatic recognition of facial actions in spontaneous expressions.J. Multimed.2006,1, 22–35. [Google Scholar]
    11. Kulkarni, S.; Reddy, N.; Hariharan, S. Facial expression (mood) recognition from facial images using committee neural networks.BioMed. Eng. OnLine2009,10, 16. [Google Scholar]
    12. Sumathi, A.; Marasamy, P. Automatic recognition and analysis of human faces and facial expression by LDA using wavelet transform.Eur. J. Sci. Res.2012,67, 389–395. [Google Scholar]
    13. Feng, X.; Pietikainen, M.; Hadid, A. Facial expression recognition with local binary patterns and linear programming.Pattern Recognition And Image Analysis C/C of Raspoznavaniye Obrazov I Analiz Izobrazhenii2005,15, 546. [Google Scholar]
    14. Kaur, M.; Vashisht, R.; Neeru, N. Recognition of facial expressions with principal component analysis and singular value decomposition.Int. J. Comput. Appl.2010,9, 24–28. [Google Scholar]
    15. Gupta, S.; Agrwal, S.; Meena, Y.; Nain, N. A Hybrid Method of Feature Extraction for Facial Expression Recognition. Proceeding of IEEE 7th International Conference on Signal-Image Technology and Internet-Based Systems, Dijon, France, 28 November–1 December 2011; pp. 422–425.
    16. Gonzalez, R.C.; Woods, R.E.Digital Image Processing; Prentice Hall: Upper Saddle River, NJ, USA, 2008. [Google Scholar]
    17. Eramian, M.; Mould, D. Histogram Equalization Using Neighborhood Metrics. Proceedings the 2nd Canadian Conference on Computer and Robot Vision, British Columbia, Canada, 9–11 May 2005; pp. 397–404.
    18. Ibrahim, H.; Kong, N.S.P. Image sharpening using sub-regions histogram equalization.IEEE Trans. Consum. Electron.2009,55, 891–895. [Google Scholar]
    19. Yeganeh, H.; Ziaei, A.; Rezaie, A. A Novel Approach for Contrast Enhancement Based on Histogram Equalization. Proceedings of International Conference on Computer and Communication Engineering, Kuala Lumpur, Malaysia, 13–15 May 2008; pp. 256–260.
    20. Shanmugavadivu, P.; Balasubramanian, K. Image inversion and Bi level histogram equalization for contrast enhancement.Int. J. Comput. Appl.2010,1, 61–65. [Google Scholar]
    21. Zuo, C.; Chen, Q.; Sui, X. Range limited Bi-histogram equalization for image contrast enhancement.Optik2013,124, 425–431. [Google Scholar]
    22. Beham, M.P.; Roomi, S.M.M. Face recognition using appearance based approach: A literature survey. Proceedings of International Conference & Workshop on Recent Trends in Technology, Mumbai, Maharashtra, India, 24–25 February 2012; pp. 16–21.
    23. Tabatabaie, Z.S.; Rahmat, R.W.; Udzir, N.I.; Kheirkhah, E. A hybrid face detection system using combination of appearance-based and feature-based methods.Int. J. Comput. Sci. Netw. Secur.2009,9, 181–185. [Google Scholar]
    24. Rivera, R.A.; Castillo, R.J.; Chae, O. Local directional number pattern for face analysis: Face and expression recognition.IEEE Trans. Image Process.2013,22, 1740–1752. [Google Scholar]
    25. Rowley, H.A.; Baluja, S.; Kanade, T. Neural network-based face detection.IEEE Trans. PAMI.1998,20, 23–38. [Google Scholar]
    26. Rosenfeld, A.; Johnshon, E. Angle detection on digital curves.IEEE Trans. Comput.1973,22, 875–878. [Google Scholar]
    27. Feng, G.C.; Yuen, P.C. Multi-cues eye detection on gray intensity image.Pattern Recognit.2001,34, 1033–1046. [Google Scholar]
    28. Bashyal, S.; Venayagamoorthy, G. Recognition of facial expressions using Gabor wavelets and learning vector quantization.Eng. Appl. Artif. Intell.2008,21, 1056–1064. [Google Scholar]
    29. Jabid, T.; Kabir, M.; Chae, O. Facial Expression Recognition Using Local Directional Pattern (LDP). Proceedings of IEEE 17th IEEE International Conference on Image Processing, Hong Kong, 26–29 September 2010; pp. 1605–1608.
    30. Rahman, A.; Ali, L. Weighted Local Directional Pattern for Robust Facial Expression Recognition. Proceedings of IEEE 2011 1st International Conference on Informatics and Computational Intelligence, Bandung, Indonesia, 12–14 December 2011; pp. 268–271.
    31. Kabir, M.; Jabid, T.; Chae, O. Local Directional Pattern Variance (LDPv): A robust feature descriptor for facial expression recognition.Int. Arab J. Inf. Technol.2012,9, 382–391. [Google Scholar]
    32. Gourier, N.; Hall, D.; Crowley, J. Estimating Face Orientation from Robust Detection of Salient Facial Structures. InFG Net Workshop on Visual Observation of Deictic Gestures; FGnet (IST–2000–26434): Cambridge, UK; 22; August; 2004; pp. 1–9. [Google Scholar]
    33. Cox, I.; Ghosn, J.; Yianilos, P. Feature-Based Face Recognition Using Mixture-Distance. 1996. Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 18–20 June 1996; pp. 209–216.
    34. Lin, D.T. Facial expression classification using PCA and hierarchical radial basis function network.J. Inf. Sci. Eng.2006,22, 1033–1046. [Google Scholar]
    35. Lin, D.T. Human Facial Expression Recognition Using Hybrid Network of PCA and RBFN. InArtificial Neural Networks–ICANN 2006; Springer: Athens, Greece, 2006; pp. 624–633. [Google Scholar]
    36. Fan, C.; Kotani, K. Facial expression recognition by supervised independent component analysis using MAP estimation.IEICE Trans. Inf. Syst.2008,91, 341–350. [Google Scholar]
    37. Buciu, I.; Pitas, I. Application of Non-Negative and Local Non Negative Matrix Factorization to Facial Expression Recognition. Proceedings of the IEEE 17th International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; Volume 1, pp. 288–291.
    38. Shinohara, Y.; Otsuf, N. Facial Expression Recognition Using fisher Weight Maps. Proceedings of Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Seoul, South Korea, 17–19 May 2004; pp. 499–504.
    39. Lyons, M.J.; Budynek, J.; Akamatsu, S. Automatic classification of single facial images.IEEE Trans. Pattern Anal. Mach. Intell.1999,21, 1357–1362. [Google Scholar]
    40. He, L.; Zhou, J.; Hu, D.; Zou, C.; Zhao, L. Boosted Independent Features for Face Expression Recognition. Proceedings of 2nd International Symposium on Neural Networks, 2005, Chongqing, China, 30 May–1 June 2005; pp. 137–146.
    41. Lirong, W.; Xiaoguang, Y.; Jianlei, W.; Jian, Z. Facial Expression Recognition Based on Local Texture Features. Proceedings of 14th IEEE International Conference on Computational Science and Engineering, Dalian, China, 24–26 August 2011; pp. 543–546.
    42. Khandait, S.P.; Thool, R.C.; Khandait, P.D. Automatic facial feature extraction and expression recognition based on neural network.Int. J. Adv. Comput. Sci. Appl.2012,2, 113–118. [Google Scholar]
    43. Pang, Y.; Yuan, Y.; Li, X. Iterative subspace analysis based on feature line distance.IEEE Trans. Image Process.2009,18, 903–907. [Google Scholar]
    44. Aguilar-Torres, G.; Toscano-Medina, K.; Perez, G.S.; Nakano-Miyatake, M.; Perez-Meana, H. Eigenface-gabor algorithm for feature extraction in face recognition.Int. J. Comput.2009,3, 20–30. [Google Scholar]
    45. Kittusamy, S.R.V.; Chakrapani, V. Facial expressions recognition using eigenspaces.J. Comput. Sci.2012,8, 1674–1679. [Google Scholar]
    46. Kalita, J.; Das, K. Recognition of facial expression using eigenvector based distributed features and euclidean distance based decision making technique.Int. J. Adv. Comput. Sci. Appl.2013,4, 196–202. [Google Scholar]
    47. Abidin, Z.; Harjoko, A. A neural network based facial expression recognition using sherface.Int. J. Comput. Appl.2012,59, 30–34. [Google Scholar]
    48. Mistry, V.J.; Goyani, M.M. A literature survey on facial expression recognition using global features.Int. J. Eng. Adv. Technol.2013,2, 653–657. [Google Scholar]
    49. Tian, Y.; Kanade, T.; Cohn, J. Facial Expression Analysis. InHandbook of Face Recognition; Springer: Heidelberg, Germany, 2005; Chapter 11; pp. 247–275. [Google Scholar]
    50. Shan, C.; Gong, S.; McOwan, P.W. Conditional Mutual Information Based Boosting for Facial Expression Recognition. InBritish Machine Vision Conference (BMVC); The British Machine Vision Association and Society for Pattern Recognition: Oxford, UK, 2005; Volume 1, pp. 399–408. [Google Scholar]
    51. Filko, D.; Goran, M. Emotion recognition system by a neural network based facial expression analysis.Autom.–J. Control Meas. Electron. Comput. Commun.2013,54, 263–272. [Google Scholar]
    52. Tu, J. Advantages and disadvantages of using artificial neural networksversus logistic regression for predicting medical outcomes.J. Clin. Epidemiol.1996,49, 1225–1231. [Google Scholar]
    53. Kotsia, I.; Pitas, I. Facial expression recognition in image sequences using geometric deformation features and support vector machines.IEEE Trans. Image Process.2007,16, 172–187. [Google Scholar]
    54. Kharat, G.; Dudul, S. Human emotion recognition system using optimally designed SVM with different facial feature extraction techniques.WSEAS Trans. Comput.2008,7, 650–659. [Google Scholar]
    55. Tsai, H.; Lai, Y.; Zhang, Y. Using SVM to Design Facial Expression Recognition for Shape and Texture Features. Proceedings of IEEE International Conference on Machine Learning and Cybernetics, Qingdao, China, 11–14 July 2010; pp. 2697–2704.
    56. Chen, F.; Kotani, K. Facial Expression Recognition by SVM-based Two-Stage Classifier on Gabor Features. Proceedings of IAPR Conference on Machine Vision Applications, Tokyo, Japan, 16–18 May 2007; pp. 453–456.
    57. Shan, C.; Gong, S.; McOwan, P. Facial expression recognition based on local binary patterns: A comprehensive study.Image Vis. Comput.2009,27, 803–816. [Google Scholar]
    58. Support Vector Machines. Available online:http://scikit-learn.org/stable/modules/svm.html (accessed on Monday 14 October 2013).
    59. Liu, W.; Lu, J.; Wang, Z.; Song, H. An Expression Space Model for Facial Expression Analysis. Proceedings of IEEE Congress on Image and Signal Processing, Sanya, China, 27–30 May 2008; Volume 2, pp. 680–684.
    60. Schels, M.; Schwenker, F. A Multiple Classifier System Approach for Facial Expressions in Image Sequences Utilizing GMM Supervectors. Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 4251–4254.
    61. Bouwmans, T.; El Baf, F. Modeling of dynamic backgrounds by type-2 fuzzy Gaussians mixture models.MASAUM J. Basic Appl. Sci.2009,1, 265–276. [Google Scholar]
    62. Yeasin, M.; Bullot, B.; Sharma, R. Recognition of facial expressions and measurement of levels of interest from video.IEEE Trans. Multimed.2006,8, 500–508. [Google Scholar]
    63. Aleksic, P.S.; Katsaggelos, A.K. Automatic facial expression recognition using facial animation parameters and multistream HMMs.IEEE Trans. Inf. Forensics Secur.2006,1, 3–11. [Google Scholar]
    64. Uddin, M.; Lee, J.; Kim, T. An enhanced independent component-based human facial expression recognition from video.IEEE Trans. Consum. Electron.2009,55, 2216–2224. [Google Scholar]
    65. Mariappan, M.B.; Suk, M.; Prabhakaran, B. Facial Expression Recognition using Dual Layer Hierarchical SVM Ensemble Classification. Proceedings of IEEE International Symposium on Multimedia, Irvine, CA, USA, 10–12 December 2012; pp. 104–107.
    66. Kim, Y.-C.; Pang, S.; Je, H.-M.; Kim, D.; Bang, S-Y. Constructing support vector machine ensemble.Pattern Recognit.2003,36, 2757–2767. [Google Scholar]
    67. Singh, S.K.; Chauhan, D.S.; Vatsa, M.; Singh, R. A robust skin color based face detection algorithm.Tamkang J. Sci. Eng.2003,6, 227–234. [Google Scholar]
    68. Hu, J.; Yan, H.; Sakalli, M. Locating head and face boundaries for head shoulder images.Pattern Recognit.1999,32, 1317–1333. [Google Scholar]
    69. Starovoitov, V.V.; Samal, D.I.; Sankur, B. Matching of Faces in Camera Images and Document Photographs. Proceedings of 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing, Istanbul, Turkey, 5–9 June 2000; pp. 2349–2352.
    70. Nusseck, M.; Cunningham, D.W.; Wallraven, C.; Bülthoff, H.H. The contribution of different facial regions to the recognition of conversational expressions.J. Vis.2008,8, 1–23. [Google Scholar]
    71. Kaulard, K.; Cunningham, D.W.; Bülthoff, H.H.; Wallraven, C. The MPI facial expression database a validated database of emotional and conversational facial expressions.PLoS One2012,7, e32321. [Google Scholar]
    72. Schmidt, K.L.; Cohn, J.F. Human facial expressions as adaptations: Evolutionary questions in facial expression research.Am. J. Phys. Anthropol.2002,116, 3–24. [Google Scholar]
    73. Kanade, T.; Cohn, J.F.; Tian, Y. Comprehensive Database for Facial Expression Analysis. Proceedings of Fourth IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France, 28–30 March 2000; pp. 46–53.
    74. Lyons, M.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding Facial Expressions with Gabor Wavelets. Proceedings of Third IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998; pp. 200–205.
    75. Samaria, F.S.; Harter, A. Parameterization of a Stochastic Model for Human Face Identification. Proceedings of the Second IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 5–7 December 1994; pp. 138–142.
    76. Georghiades, A.S.; Belhumeur, P.N.; Kriegman, D.J. From few to many: Illumination cone models for face recognition under variable lighting and pose.IEEE Trans. Pattern Anal. Mach. Intell.2001,23, 643–660. [Google Scholar]
    77. Shamir, L. Evaluation of face datasets as tools for assessing the performance of face recognition methods.Int. J. Comput. Vis.2008,79, 225–230. [Google Scholar]
    78. Zhou, J.; Xu, T.; Gan, J. Facial Expression Recognition Based on Local Directional Pattern Using SVM Decision-Level Fusion. Proceedings of Tenth International Conference on Computability and Complexity in Analysis, Nancy, France, 8–10 July 2013; pp. 126–132.
    79. Zhao, X.; Zhang, S. Facial expression recognition based on local binary patterns and kernel discriminant isomap.Sensors2011,11, 9573–9588. [Google Scholar]
    Sensors 13 16682f1 1024
    Figure 1. Typical architecture of general facial expression recognition systems.
    Figure 1. Typical architecture of general facial expression recognition systems.
    Sensors 13 16682f1 1024
    Sensors 13 16682f2 1024
    Figure 2. 3D-feature plot of Cohn-Kanade dataset for six different types of facial expressions, where each expression has twelve expression frames. It can be seen that the features are highly merged, due to the presence of similarity between the expressions, which could later results in a high misclassification rate.
    Figure 2. 3D-feature plot of Cohn-Kanade dataset for six different types of facial expressions, where each expression has twelve expression frames. It can be seen that the features are highly merged, due to the presence of similarity between the expressions, which could later results in a high misclassification rate.
    Sensors 13 16682f2 1024
    Sensors 13 16682f3 1024
    Figure 3. Architectural diagram for the HL-FER.
    Figure 3. Architectural diagram for the HL-FER.
    Sensors 13 16682f3 1024
    Sensors 13 16682f4 1024
    Figure 4. The control flow diagram for the proposed face detection system.
    Figure 4. The control flow diagram for the proposed face detection system.
    Sensors 13 16682f4 1024
    Sensors 13 16682f5 1024
    Figure 5. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on Cohn-Kanade dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Figure 5. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on Cohn-Kanade dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Sensors 13 16682f5 1024
    Sensors 13 16682f6 1024
    Figure 6. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on Cohn-Kanade dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 7.
    Figure 6. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on Cohn-Kanade dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 7.
    Sensors 13 16682f6 1024
    Sensors 13 16682f7 1024
    Figure 7. 3D-feature plot of single-LDA via single-HMM (on Cohn-Kanade dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 6).
    Figure 7. 3D-feature plot of single-LDA via single-HMM (on Cohn-Kanade dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 6).
    Sensors 13 16682f7 1024
    Sensors 13 16682f8 1024
    Figure 8. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on JAFFE dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Figure 8. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on JAFFE dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Sensors 13 16682f8 1024
    Sensors 13 16682f9 1024
    Figure 9. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on JAFFE dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 10.
    Figure 9. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on JAFFE dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 10.
    Sensors 13 16682f9 1024
    Sensors 13 16682f10 1024
    Figure 10. 3D-feature plot of single-LDA via single-HMM (on JAFFE dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 9).
    Figure 10. 3D-feature plot of single-LDA via single-HMM (on JAFFE dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 9).
    Sensors 13 16682f10 1024
    Sensors 13 16682f11 1024
    Figure 11. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on AT&T dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Figure 11. 3D feature plots of the HL-FER after applying LDA at the first level for the three expression-categories such as lips-based, lips-eyes-based, or lips-eyes-forehead-based expressions (on AT&T dataset). It can be seen that at the first level, the HL-FER achieved 100% classification rate in expressions categories classification.
    Sensors 13 16682f11 1024
    Sensors 13 16682f12 1024
    Figure 12. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on AT&T dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 13.
    Figure 12. 3D feature plots of the HL-FER after applying LDA at the second level for recognizing the expressions in each category (on AT&T dataset). It can be seen that at the second level, the HL-FER achieved much higher recognition rate as compared to a single-LDA via single-HMM shown inFigure 13.
    Sensors 13 16682f12 1024
    Sensors 13 16682f13 1024
    Figure 13. 3D-feature plot of single-LDA via single-HMM (on AT&T dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 12).
    Figure 13. 3D-feature plot of single-LDA via single-HMM (on AT&T dataset). It can be seen that using a single-LDA via single-HMM approach did not yield as good a separation among different classes as was achieved by the HL-FER (SeeFigure 12).
    Sensors 13 16682f13 1024
    Sensors 13 16682f14 1024
    Figure 14. Sample results for the proposed GHE along with the face detection method.
    Figure 14. Sample results for the proposed GHE along with the face detection method.
    Sensors 13 16682f14 1024
    Sensors 13 16682f15 1024
    Figure 15. Sample reference images for three datasets: from left to right: Cohn-Kanade dataset, AT&T dataset, and JAFFE dataset, respectively.
    Figure 15. Sample reference images for three datasets: from left to right: Cohn-Kanade dataset, AT&T dataset, and JAFFE dataset, respectively.
    Sensors 13 16682f15 1024
    Table
    Table 1. The classified categories and facial expressions recognized in this study.
    Table 1. The classified categories and facial expressions recognized in this study.
    CategoryFacial Expressions
    Lips-BasedHappiness
    Sadness
    Lips-Eyes-BasedSurprise
    Disgust
    Lips-Eyes-Forehead-BasedAnger
    Fear
    Table
    Table 2. Average accuracy rates of the proposed face detection algorithm on three publically available datasets (Unit: %).
    Table 2. Average accuracy rates of the proposed face detection algorithm on three publically available datasets (Unit: %).
    DatasetsDetection Rate
    Cohn-Kanade99
    JAFFE99
    AT&T98
    Average98.7
    Table
    Table 3. Confusion matrix for the HL-FER using the Cohn-Kanade database of facial expressions (Unit: %).
    Table 3. Confusion matrix for the HL-FER using the Cohn-Kanade database of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9820000
    Sadness2980000
    Anger0099001
    Disgust0009910
    Surprise00001000
    Fear0020098
    Average98.87
    Table
    Table 4. Confusion matrix for the HL-FER using the JAFFE database of facial expressions (Unit: %).
    Table 4. Confusion matrix for the HL-FER using the JAFFE database of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9910000
    Sadness2980000
    Anger00100000
    Disgust0009910
    Surprise0003970
    Fear00000100
    Average98.80
    Table
    Table 5. Confusion matrix for the HL-FER using the AT&T database of facial expressions (Unit: %).
    Table 5. Confusion matrix for the HL-FER using the AT&T database of facial expressions (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9820000
    Sadness1990000
    Anger0099001
    Disgust0009820
    Surprise0002980
    Fear0010099
    Average98.50
    Table
    Table 6. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on Cohn-Kanade dataset (Unit: %).
    Table 6. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on Cohn-Kanade dataset (Unit: %).
    Expressions CategoryRecognition Rate
    Lips-Based86
    Lips-Eyes-Based83
    Lips-Eyes-Forehead-Based87
    Average85.3
    Table
    Table 7. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on the JAFFE dataset (Unit: %).
    Table 7. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on the JAFFE dataset (Unit: %).
    Expressions CategoryRecognition Rate
    Lips-Based80
    Lips-Eyes-Based79
    Lips-Eyes-Forehead-Based82
    Average80.3
    Table
    Table 8. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on the AT&T dataset (Unit: %).
    Table 8. The recognition rate of the HL-FER at the first level for recognizing the expressions category such as Lips-Based, Lips-Eyes-Based, and Lips-Eyes-Forehead-Based on the AT&T dataset (Unit: %).
    Expressions CategoryRecognition Rate
    Lips-Based87
    Lips-Eyes-Based84
    Lips-Eyes-Forehead-Based82
    Average84.3
    Table
    Table 9. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on Cohn-Kanade dataset and testing on the JAFFE and AT&T datasets (Unit: %).
    Table 9. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on Cohn-Kanade dataset and testing on the JAFFE and AT&T datasets (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness8143426
    Sadness5852314
    Anger2382535
    Disgust5437576
    Surprise2346805
    Fear2431585
    Average81.30
    Table
    Table 10. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on the JAFFE dataset and testing on the Cohn-Kanade and AT&T datasets (Unit: %).
    Table 10. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on the JAFFE dataset and testing on the Cohn-Kanade and AT&T datasets (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness7147657
    Sadness8746534
    Anger4679533
    Disgust3248173
    Surprise5542804
    Fear3582676
    Average76.80
    Table
    Table 11. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on the AT&T dataset and testing on the Cohn-Kanade and JAFFE datasets (Unit: %).
    Table 11. Confusion matrix for the HL-FER, showing the weighted average recognition accuracy for six expressions. Training on the AT&T dataset and testing on the Cohn-Kanade and JAFFE datasets (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness8064235
    Sadness7795414
    Anger4677535
    Disgust5587246
    Surprise2354833
    Fear2560384
    Average79.20
    Table
    Table 12. Confusion matrix for the HL-FER (on the Cohn-Kanade dataset), while removing the preprocessing step (Unit: %).
    Table 12. Confusion matrix for the HL-FER (on the Cohn-Kanade dataset), while removing the preprocessing step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9280000
    Sadness6940000
    Anger0095005
    Disgust0009370
    Surprise00010900
    Fear0060094
    Average93.00
    Table
    Table 13. Confusion matrix for the HL-FER (on the Cohn-Kanade dataset), while removing the hierarchical recognition step (Unit: %).
    Table 13. Confusion matrix for the HL-FER (on the Cohn-Kanade dataset), while removing the hierarchical recognition step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness8920045
    Sadness0924400
    Anger0590500
    Disgust00118900
    Surprise4006900
    Fear0290089
    Average89.80
    Table
    Table 14. Confusion matrix for the HL-FER (on the JAFFE dataset), while removing the preprocessing step (Unit: %).
    Table 14. Confusion matrix for the HL-FER (on the JAFFE dataset), while removing the preprocessing step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9190000
    Sadness7930000
    Anger0096004
    Disgust00090100
    Surprise0008920
    Fear0080092
    Average92.33
    Table
    Table 15. Confusion matrix for the HL-FER (on the JAFFE dataset), while removing the hierarchical recognition step (Unit: %).
    Table 15. Confusion matrix for the HL-FER (on the JAFFE dataset), while removing the hierarchical recognition step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness8253253
    Sadness0903232
    Anger1393201
    Disgust4068703
    Surprise2400886
    Fear2067085
    Average87.50
    Table
    Table 16. Confusion matrix for the HL-FER (on the AT&T dataset), while removing the preprocessing step (Unit: %).
    Table 16. Confusion matrix for the HL-FER (on the AT&T dataset), while removing the preprocessing step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness89110000
    Sadness9910000
    Anger0096004
    Disgust00090100
    Surprise00012880
    Fear00100090
    Average90.66
    Table
    Table 17. Confusion matrix for the HL-FER (on the AT&T dataset), while removing the hierarchical recognition step (Unit: %).
    Table 17. Confusion matrix for the HL-FER (on the AT&T dataset), while removing the hierarchical recognition step (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9203203
    Sadness3890233
    Anger2488024
    Disgust5248522
    Surprise2304910
    Fear1408384
    Average88.16
    Table
    Table 18. Confusion matrix of the HL-FER trained on the Cohn-Kanade dataset and tested on the AT&T dataset (Unit: %).
    Table 18. Confusion matrix of the HL-FER trained on the Cohn-Kanade dataset and tested on the AT&T dataset (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness9142210
    Sadness4903003
    Anger1292005
    Disgust3138940
    Surprise0233902
    Fear0231292
    Average90.70
    Table
    Table 19. Confusion matrix of the HL-FER trained on the AT&T dataset and tested on the Cohn-Kanade dataset (Unit: %).
    Table 19. Confusion matrix of the HL-FER trained on the AT&T dataset and tested on the Cohn-Kanade dataset (Unit: %).
    HappinessSadnessAngerDisgustSurpriseFear
    Happiness8942212
    Sadness2920240
    Anger0388423
    Disgust0249022
    Surprise5123872
    Fear1312390
    Average89.30
    Table
    Table 20. Comparison results of the HL-FER with some of the existing works for the three standard Cohn-Kanade, JAFFE, and AT&T datasets for six basic facial expression under the same settings as described in Section 3.1 (Unit: %).
    Table 20. Comparison results of the HL-FER with some of the existing works for the three standard Cohn-Kanade, JAFFE, and AT&T datasets for six basic facial expression under the same settings as described in Section 3.1 (Unit: %).
    Existing Work[64][34][35][36][37][38][39][40][79]HL-FER
    Average Accuracy Rate91868785816982868798

    © 2013 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).

    Share and Cite

    MDPI and ACS Style

    Siddiqi, M.H.; Lee, S.; Lee, Y.-K.; Khan, A.M.; Truc, P.T.H. Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems.Sensors2013,13, 16682-16713. https://doi.org/10.3390/s131216682

    AMA Style

    Siddiqi MH, Lee S, Lee Y-K, Khan AM, Truc PTH. Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems.Sensors. 2013; 13(12):16682-16713. https://doi.org/10.3390/s131216682

    Chicago/Turabian Style

    Siddiqi, Muhammad Hameed, Sungyoung Lee, Young-Koo Lee, Adil Mehmood Khan, and Phan Tran Ho Truc. 2013. "Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems"Sensors 13, no. 12: 16682-16713. https://doi.org/10.3390/s131216682

    APA Style

    Siddiqi, M. H., Lee, S., Lee, Y.-K., Khan, A. M., & Truc, P. T. H. (2013). Hierarchical Recognition Scheme for Human Facial Expression Recognition Systems.Sensors,13(12), 16682-16713. https://doi.org/10.3390/s131216682

    Article Metrics

    No
    No

    Article Access Statistics

    For more information on the journal statistics, clickhere.
    Multiple requests from the same IP address are counted as one view.
    Sensors, EISSN 1424-8220, Published by MDPI
    RSSContent Alert

    Further Information

    Article Processing Charges Pay an Invoice Open Access Policy Contact MDPI Jobs at MDPI

    Guidelines

    For Authors For Reviewers For Editors For Librarians For Publishers For Societies For Conference Organizers

    MDPI Initiatives

    Sciforum MDPI Books Preprints.org Scilit SciProfiles Encyclopedia JAMS Proceedings Series

    Follow MDPI

    LinkedIn Facebook X
    MDPI

    Subscribe to receive issue release notifications and newsletters from MDPI journals

    © 1996-2025 MDPI (Basel, Switzerland) unless otherwise stated
    Terms and Conditions Privacy Policy
    We use cookies on our website to ensure you get the best experience.
    Read more about our cookieshere.
    Accept
    Back to TopTop
    [8]ページ先頭

    ©2009-2025 Movatter.jp