999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Face Recognition Based on Gabor Feature Extraction Followed by FastICA and LDA

2021-12-11 13:29:28MasoudMuhammedHassanHavalIsmaelHusseinAdelSabryEesaandRamadhanMstafa
Computers Materials&Continua 2021年8期

Masoud Muhammed Hassan,Haval Ismael Hussein,Adel Sabry Eesa and Ramadhan J.Mstafa,2

1Department of Computer Science,Faculty of Science,University of Zakho,Duhok,42002,Kurdistan Region,Iraq

2Scientific Research and Development Center,Nawroz University,Duhok,42001,Kurdistan Region,Iraq

Abstract: Over the past few decades, face recognition has become the most effective biometric technique in recognizing people’s identity, as it is widely used in many areas of our daily lives.However,it is a challenging technique since facial images vary in rotations,expressions, and illuminations.To minimize the impact of these challenges, exploiting information from various feature extraction methods is recommended since one of the most critical tasks in face recognition system is the extraction of facial features.Therefore, this paper presents a new approach to face recognition based on the fusion of Gabor-based feature extraction,Fast Independent Component Analysis(FastICA), and Linear Discriminant Analysis (LDA).In the presented method,first,face images are transformed to grayscale and resized to have a uniform size.After that,facial features are extracted from the aligned face image using Gabor,FastICA,and LDA methods.Finally,the nearest distance classifier is utilized to recognize the identity of the individuals.Here,the performance of six distance classifiers,namely Euclidean,Cosine,Bray-Curtis,Mahalanobis,Correlation,and Manhattan,are investigated.Experimental results revealed thatthe presented method attainsa higher rank-one recognition rate compared to the recent approaches in the literature on four benchmarked face datasets:ORL,GT,FEI,and Yale.Moreover,it showed that the proposed method not only helps in better extracting the features but also in improving the overall efficiency of the facial recognition system.

Keywords: Artificial intelligence; face recognition; FastICA; Gabor filter bank; LDA

1 Introduction

Recently, biometric-adopted techniques have become the most promising option for recognizing people.Among many biometric technologies, face recognition has become the main focus of biometrics research, as it is widely employed in many fields, for instance, computer systems, artificial intelligence, access control for security facilities, automobiles, image analysis, and authentication processes.The reason why it is so widespread is that facial images can be easily accumulated without any pre-approval and can also be effortlessly obtained from a distance [1].Although it has been an ongoing field of research since the 1960s, it is still constantly developing due to several challenges encountering the recognition systems [2,3].These challenges are changes in illumination, facial hairs (e.g., eyebrows, beard, head hairstyle, and mustache), expressions(e.g., sadness, anger, disgust, surprise, fear, and happiness), and wearing of accessories (e.g., hats,scarf, and glasses).Because of these differences in image quality, there might be a likelihood of low variance between image samples of different individuals and high variance between image samples of the same individual.This similarity in inter-class images and dissimilarity in intraclass images can significantly deteriorate the efficacy of the recognition system [4].Therefore,many researchers have innovated new algorithms or developed the existing ones to overcome the problems mentioned above.

In general, face recognition system consists of four consecutive phases:face detection, face alignment, feature extraction, and face recognition (see Fig.1).Face detection refers to the process of identifying human faces in an image or video.The next step is face alignment, where some pre-processing steps are applied to the detected face images, such as transforming face images to grayscale, uniformity of facial images size.After that, the feature extraction step is performed,where facial features are extracted, such as the distance between eyes, nose width, depth of eye socket, and length of the jawline.Lastly, the face recognition system correlates the features of a person’s image with the features of all enrolled people in the database, and then the person is recognized [5].

Figure 1:Generic diagram of face recognition system [6]

Face recognition methods can be categorized into four techniques, as follows:part based technique, holistic based technique, template based technique, and feature based technique [7,8].Part based technique identifies the vital part of facial images, and then merge artificial intelligence and machine learning tools with the part appearances for recognition.A component-based approach such as Support Vector Machine (SVM) is a part-based technique.Holistic based technique takes the entire face area as input data for the facial recognition system.Independent Component Analysis (ICA), Linear Discriminant Analysis (LDA), and Principal Component Analysis (PCA)are some methods of the holistic based technique.Template based technique compares a collection of stored templates of the faces with the input face image.Templates can be built by employing tools such as LDA, SVM, and PCA.Active Shape Model and Elastic Bunch Graph Matching are some methods of the template-based technique.Feature based technique makes use of computer vision and image processing techniques to extract features from the entire face or specific points of the face.Local Binary Pattern (LBP) and Gabor wavelet features are some methods of the feature based technique [4,8].

In this paper, a new feature extraction technique for face recognition is proposed based on the combination of Gabor-based feature extraction, FastICA, and LDA methods.The motivation behind choosing these three methods as a new fusion approach is to provide a more accurate face recognition method under different challenging conditions.As highlighted by [9], the key advantage of using Gabor filters is their invariance to pose variations, as well as their robustness against image noise and illumination variations [10].On the other hand, FastICA and LDA are the most successful methods to find independent components and to best discriminates different classes.They provide discriminated and independent features which are robust to the facial variations as reported by [11,12].Four face datasets:GT, Yale, FEI, and ORL are used to assess the efficiency of the proposed method under conditions where face images are varied in illumination, facial expressions, and rotations.Moreover, six well-known distance measures, namely Euclidean, Cosine,Bray-Curtis, Mahalanobis, Correlation, and Manhattan, are investigated for face classification.

The rest of the paper is structured as follows.Section 2 reviews some related recent approaches to face recognition.Section 3 provides a brief theoretical background on Gabor-based feature extraction, FastICA, and LDA methods.Besides, it presents the mathematical formulas of the six dissimilarity measures used.Section 4 elucidates the proposed face recognition method.Section 5 exhibits the experimental results.Finally, Section 6 concludes the paper and suggests some future works.

2 Literature Survey

Face recognition is still an area of active research since facial images vary in rotations,expressions, lighting, and occlusions.There are different approaches to tackle some of the above issues.One of them is to extract the most distinctive features from facial images by exploiting the merits of several feature extraction methods.In our literature survey, we focused on the recent techniques of face recognition based on a single feature extraction method or a combination of more than one method.Various approaches have been suggested in recent years to handle the challenges arisen from the face recognition system.

Aldhahab et al.[13] presented a technique based on the fusion of Discrete Cosine Transform(DCT) and Vector quantization (VQ) for face recognition.DCT method was employed to extract features from the face image.Whereas, VQ method was used to reduce the dimensionality of the extracted features.For face classification, they used Euclidean distance measure.Their experimental results exhibited an improvement in recognition rates on four face datasets:FEI, YALE, ORL,and FERET.

Aldhahab et al.[14] proposed a new method to face recognition based on Two-Dimensional Discrete Multiwavelet Transform (2D-DMWT).In their method, the face image was first partitioned into six portions, and then 2D-DMWT was applied on each portion to reduce the dimensionality of facial features.Lastly, for more data compression, L2 norm was applied.They adopted Neural Network (NN) for face classification.Their method was tested on four face datasets:FERET, FEI, YALE, and ORL.

Muqeet et al.[15] developed a new feature extraction technique for face recognition using local binary patterns (LBP) and interpolation-based directional wavelet transform (DIWT).In the implementation of DIWT, first, quadtree partitioning was employed to form the sub-bands for the directional wavelet transform.Then LBP was performed to the chosen top-level DIWT sub-bands,which ultimately extract the descriptive histogram features.Lastly, LDA was performed to get discriminant features in reduced space.In the identification process, the Nearest Neighbor (NN)classifier was employed.Three different face datasets, namely ORL, GT, and FEI, were used to test their presented technique.

Dora et al.[16] presented an Evolutionary Single Gabor Kernel (ESGK) approach based on the fusion of two bio-inspired optimization methods to overcome the computational complexity of Gabor filter banks in face recognition.These two bio-inspired methods are Particle Swarm Optimization (PSO) and Gravitational Search Algorithm (GSA), where their fusion is called PSOGSA.The hybrid PSO-GSA was utilized to optimize the parameters of the single Gabor filter and to extract features from facial images.They also presented a new classification method based on the eigenvalue approach for face recognition.Their presented approach was assessed on five face datasets:LFW, ORL, GT, UMIST, and FERET.Experimental results confirmed the efficiency of their new proposed method compared to the conventional Gabor filter bank method.

Wei et al.[17] presented a new method to face recognition based on fuzzy set and LDA,called Complete Fuzzy LDA (CFLDA) method, where fuzzy set theory was fully integrated into LDA model.They used the nearest distance classifier for face image classification.A collection of experiments was conducted to validate the performance of their presented method on three face datasets:ORL, FERET, and Yale.

Liao et al.[18] presented two new methods for face recognition, namely logarithmic weighted sum (LWS) and extended sparse weighted representation classifier (ESWRC).LWS was used to extract facial features.Whereas, ESWRC was employed for face classification.Experimental results revealed that their presented approach performs better than the recent approaches on FERET and FEI face datasets.

Fan et al.[19] proposed an effective face recognition algorithm called joint collaborative representation, which jointly uses space-domain features and frequency-domain features.For face classification, they applied two distance-based classifiers to two sets of data (space-domain features and frequency-domain features), where they employed an adaptive weight to integrate them.Empirical results showed that their proposed method attains better results than the traditional collaborative representation method on three face datasets, namely FERET, GT, and ORL.

Ouyang et al.[20] introduced a new approach to face recognition that integrates Improved Kernel Linear Discriminant Analysis (IKLDA) algorithm with Probabilistic Neural Networks(PNNs) algorithm.IKLDA algorithm was used to reduce the dimensions of face images.For face classification, PNN algorithm was used.A collection of experiments on AR, ORL, and Yale face datasets was carried out to assess the feasibility of their presented method.Their results indicated that IKLDA+PNN method can improve both the computational efficiency and face recognition performance.

Ayyad et al.[21] presented a hybrid approach to face recognition based on the fusion of Singular Value Decomposition (SVD) and Relevance Weighted Linear Discriminant Analysis (RWLDA) in Discrete wavelet transform domain.Their fusion method was employed to extract the significant features from facial images.They used the nearest neighbor classifier for face classification.ORL and GT face datasets were used to conduct the experiments.They concluded that the use of Z-score and Min-Max normalization techniques along with their proposed fusion method can improve the training time and recognition rate.

Song et al.[22] proposed a new approach to face recognition based on the fusion of local and global Gaussian Hermite Moments (GHMs).To extract both the local and global features of the face image, they firstly used GHMs as a facial feature, where they could construct the spatial pyramid of the face image.Then, they computed the scatter-ratio to choose the most distinctive features.Finally, for face classification, they employed Sparse Representation Classifier (SRC).A set of experiments were performed on Yale, FERET, and ORL face datasets to measure the efficiency of their proposed method.Experimental results showed that their proposed method outperforms other related fusion methods, such as (PCA+SVM), (Gabor+SRC), and(LDA+SRC).

Gou et al.[23] presented two new linear reconstruction classification techniques based upon Sparsity Augmented Collaborative Representation Classification technique (SACRC) to improve the representation classification performance.The first proposed technique is the Weighted Enhancement Linear Reconstruction Measure Classification technique (WELRMC) that incorporates data localities into SACRC.Whereas, the second proposed technique is the two-phase Weighted Enhancement Linear Reconstruction Measure Classification (TP-WELRMC) that fuses both the fine and the coarse representations into SACRC.Empirical results demonstrated that their proposed method obtains better results than the recent RBC approaches on five face datasets:LFW, IMM, ORL, GT, and Yale.

Liao et al.[24] proposed a face recognition approach based on subspace extended sparse representation and learning discriminative feature, named SESRC&LDF.In their method, each test image is deemed to have small or large pose variation.For face classification, if the test image is deemed to be the earlier, SESRC will be used.Otherwise, LDF will be employed.Empirical results demonstrated that their approach attains the highest recognition rates compared to many state-of-the-art approaches on eight face datasets:Georgia Tech, Extended Yale B, AR, Yale, FEI,LFW, UMIST, and FERET.

Based on our literature survey and to the best of our knowledge, none of the previous works has investigated the combination of Gabor feature extraction with two well-known dimensionality reduction methods, namely FastICA and LDA.Motivated by the successful use of Gabor filters in different fields, including computer vision and face recognition, it would be of great interest to examine the combination of Gabor filters with FastICA and LDA in face recognition problem.Therefore, we aim to extract features from the facial image using Gabor filter bank in the first phase, then reduce its features using FastICA method, and lastly, reduce FastICA features using LDA method.The proposed method will be evaluated on Yale, ORL, FEI, and GT face datasets.Furthermore, the performance of the proposed method will be investigated using six distance measures, e.g., Euclidean, Cosine, Bray-Curtis, Mahalanobis, Correlation, and Manhattan.

3 Background

This section reviews the techniques of face feature extraction used in the proposed method.Furthermore, it shows the mathematical background of the six dissimilarity measures used in the recognition phase.

3.1 Feature Extraction Based on Gabor Filter Bank

Gabor filters are bio-inspired convolutional kernels that have extensive use in image processing and computer vision fields [25].These filters have two interesting characteristics:selectivity for orientation and location frequency, which are analogous to the human visual system [26].Therefore, analyzing images using these filters are found out to be particularly useful for texture discrimination and representation [27].In the spatial domain, a 2-D Gabor filter is a Gaussian kernel method formed by a complex sinusoidal plane wave, which is defined as in Eq.(1) [28]:

wherefindicates the sine frequency,θindicates the orientation of the Gabor method,φrepresents the phase offset,αandβare the Gaussian sharpness with the major axis parallel and the minor axis perpendicular to the wave, respectively.σindicates the standard deviation of the Gaussian envelope, and γγ indicates the spatial aspect ratio.For more information about how the Gabor filter bank is used, the reader is referred to [28].

Gabor filters with different orientations and scales can be employed to generate a Gabor filter bank.Mostly, a Gabor filter bank with eight orientations and five scales is used (see Fig.2).These filters are used to extract features from grayscale images.The process of Gabor-based feature extraction works as follows:at first, each filter in the Gabor filter bank is convoluted with the image.Then, the resulting filtered images are down-sampled to reduce the redundant information.Next, each down-sampled image is transformed into a feature vector.After that, each feature vector is normalized to zero mean and unit variance.Finally, the normalized feature vectors are combined to produce the final feature vector of the image [9].

Figure 2:Gabor filters with eight orientations and five scales [9]

3.2 Fast Independent Component Analysis(FastICA)

FastICA is a faster and effective method for ICA, which is first suggested and invented by [29,30].There are a few interesting points in employing FastICA rather than ICA algorithm.For instance, FastICA converges faster and requires no step size determination as compared to ICA based on a gradient-based method.FastICA obtains an independent Non-Gaussian feature by any random nonlinear functionθ, whereas other methods require assessing the Probability Density Function (PDF) of the specified nonlinear functions.

FastICA method attempts to look for a local extreme of Kurtosis for the observed blending signals.The objective of the fixed point is to search for the maximum ofoi=w′iTx′which has maximum non-Gaussianity.FastICA utilizes negentropy as the objective function [12].Reference [31] presented an easy way to approximate the negentropy function as in Eq.(2).

whereoirepresents the output random variable with unit variance and zero mean,G′represents a nonquadratic function, andvrepresents a Gaussian random variable with unit variance and zero mean.The approximation given in Eq.(3) leads to the use of a new objective function for estimating ICA.Thus, maximizing the functionJ′G′gives one independent component.In Eq.(3),w′represents thep-dimensional weighted vectors.The method for FastICA based on maximum negentropy objective function is elucidated in detail in [31].

3.3 Linear Discriminant Analysis(LDA)

LDA is a supervised linear dimensionality reduction method that looks for the subspace that best discriminates different classes.The primary purpose of LDA is to apply dimensionality reduction while keeping up as much class discriminatory information as possible [32].It transforms a high-dimensional feature vector to a low-dimensional space so that the ratio between the interclass scatterSBand the intra-class scatterSWis maximized.For the multi-class classification problem that containscclass labels, these scatters can be defined as in Eqs.(4) and (5) [25,33]:

whererepresents theith sample of classj,μjrepresents the mean of classj,crepresents the number of classes,Njrepresents the number of training samples in classj, andμrepresents the total mean of whole classes.SBrefers to the scatter of features around the total mean of whole classes, andSWrefers to the scatter of features around the mean of each class [33,34].

LDA aims to minimizeSWand meanwhile maximizeSB.To achieve this, the ratio of|SB|/|SW|has to be maximized.This ratio is maximized when the column vectors of the projection matrixWare the eigenvectors ofS?1W SW.To preventSWfrom being singular, PCA is employed as a pre-processing step of LDA.Therefore, the originals-dimensional space is projected onto a mediumt-dimensional space by PCA, then lastly onto anf-dimensional space by LDA(f?t?s) [33].

3.4 Dissimilarity Measures

Distance measures are employed to calculate the ratio of similarity or dissimilarity between two lists of numbers (for instance, vectors).There are over sixty distinct similarity measures,many of which are used to recognize faces [35].In this paper, the proposed method investigates the performance of six well-known distance measures under conditions where face images vary in rotations, facial expressions, and illuminations.These distance measures are Euclidean [36],Cosine [37], Bray-Curtis [33], Mahalanobis [38], Correlation [38], and Manhattan [38].The mathematical formulas of the six distance measures are presented in Eqs.(6)-(11), respectively.In the following equations,nrepresents the dimensionality of the face image,xrepresents the test face image,yrepresents the training face image, andxiandyirepresent theith values ofxandyface images to be compared.

whereS?1represents the inverse of the covariance matrix betweenxandy.

whereCovis the covariance andσxandσyare the standard deviations ofxandy.

4 The Proposed Face Recognition Method

In this section, we present our proposed approach to face recognition based on the fusion of Gabor features, FastICA, and LDA.Fig.3 shows the block diagram of the proposed method in which consists of two main phases:training phase and recognition (testing) phase.

Figure 3:Block diagram of the proposed method

In the training phase, each face image was subject to the face alignment step, where face images were transformed to grayscale and resized to 64×64 pixels to have a uniform size.The reason for selecting the size of 64×64 pixels was based on independent empirical tests in which we obtained a higher rank-one recognition rate than other possible sizes.After that, Gabor-based feature extraction was used to extract features from the aligned face images.

Here is how we used Gabor-based feature extraction in the proposed method.We first generated a Gabor filter bank with eight orientations and five scales, as shown in Fig.2 presented in Section 3.1.Then, we convoluted each filter, in the Gabor filter bank, with the aligned face image.Fig.4 shows the results of a face image from Yale face dataset after being convoluted with forty filters depicted in Fig.2.Since the contiguous pixels in an image are closely related in most cases, the redundant information can be reduced by down-sampling the images originating from Gabor filters [9].Hence, each convoluted image was down-sampled by a factor of four.The reason for choosing factor four was based on empirical tests that have been conducted in our experiments.After that, each down-sampled image was transformed into a feature vector.Then,Z-score normalization was applied to each feature vector as in Eq.(12).Finally, the normalized feature vectors were merged to form the final feature vector that represents the face image.

whereFrepresents the feature vector to be normalized,μandσrepresent its mean and standard deviation, respectively.

Figure 4:Results of convoluting 40 filters depicted in Fig.2 on a face image

Since Gabor-based feature extraction results in a huge set of facial features, we employed FastICA algorithm to further reduce its dimensionality.FastICA is the most successful method to find independent components.We first found a local extreme of Kurtosis for the extracted features from the Gabor-based method in the previous step.Then, we maximized the negentropy objective function using Eq.(3) to search for the maximum ofoi=′.After that, the approximation given in Eq.(3) was used to estimate components as independent as possible, as well as the projection matrixw′which has ther-dimensional weighted vector.Thus, the number of desired components were found.

Subsequently, we used LDA algorithm to extract the most discriminant features extracted from FastICA to separate different classes as follows.First, thed-dimensional mean vectors for each classmcwas calculated.After that, the inter-class scatterSBand the intra-class scatterSWwere computed using Eqs.(4) and (5), respectively.Then, the eigenvalues(λ1,λ2,...,λd)and their corresponding eigenvectors(e1,e2,...,ed)were computed from the matrix obtained bySW?1SB.The eigenvectors were then sorted in descending order based on the eigenvalues.To form ad×kprojection matrixW, we chosekeigenvectors with the largest eigenvalues.The obtained(d×k)eigenvector matrix was then used to transform the samples onto the new subspace by the matrix multiplication (Y=X×W), whereXis an×ddimensional matrix representing thensamples,andyis the transformedn×kdimensional samples in the new subspace.Finally, all the training feature vectors with their labels were stored in a database to be used for face recognition.

In the recognition phase, a feature vector was extracted for each test face image as performed in the training phase.Then, the distance between each test feature vector and all the training feature vectors, stored in the database, was computed.From the distance measure, the face ID of the training feature vectors with the minimum distance represents the identity of the test face image.Here, we investigated the performance of six distance measures, e.g., Euclidean, Cosine,Bray-Curtis, Mahalanobis, Correlation, and Manhattan, to find out which one of them works better with the proposed (Gabor+FastICA+LDA) feature extraction method.The proposed face recognition method is shown in detail in Algorithms 1 and 2.

Algorithm 1:The proposed face recognition method Required Input:Training and testing sample sets: Xtrm×n and Xtesm×n.Required Output:Label vector yi ∈{1,2,3,...,C)(1) Sample alignment:Transform face images into grayscale [0-255] and resize it to 64 × 64 pixels.(2) Split face images into training and testing samples.(3) Ge nerate a Gabor filter bank with 8 orientations and 5 scales.(4) Use Algorithm 2 to extract features for each training face image.(5) Store feature vectors of all the training face samples along with their IDs into a database.(6) Face recognition step:(a) For each test face image, extract a feature vector using Algorithm 2.(b) Apply distance measure:(i) Calculate the distance between each test feature vector and all the training feature vectors stored in the database.(ii) The test face image is then classified into the class which has the minimum distance.(c) Identify the test face ID.(7) End

Algorithm 2:The proposed (Gabor+FastICA+LDA) feature extraction method Required Input:Aligned face image.Required Output:Feature vector(1) Apply Gabor feature extraction:(a) Convolute each Gabor filter with the aligned face image.(b) Down-sample each convoluted image by a factor of 4.(c) Transform each down-sampled image into a feature vector.(d) Apply Z-score normalization to each feature vector.(e) Merge the normalized features to form the final feature vector.(2) Apply FastICA algorithm to reduce the dimensionality of Gabor features:(a) Find a local extreme of Kurtosis for the extracted features from Step 1.(b) Maximize negentropy objective function using Eq.(3).(c) Find k:the number of desired components.(3) Apply LDA to extract the most discriminant features from the FastICA features:(a) Calculate the d-dimensional mean vectors for each class, mc.(b) Compute inter-class scatter SB and intra-class scatter SW using Eqs.(4) and(5), respectively.(c) Compute eigenvalues (λ1,λ2,...,λd) and corresponding eigenvectors (e1,e2,...,ed)from SW ?1SB.(d) Sort the eigenvectors in descending order and choose k eigenvectors with the largest eigenvalues to form a d×k projection matrix W.(e) Use this (d×k) eigenvector matrix to transform the samples onto the new subspace by the matrix multiplication: Y=X×W.(4) End?

5 Experimental Results

To evaluate the efficiency of the proposed method, a set of experiments were carried out on four face datasets, namely GT, ORL, Yale, and FEI.These four face datasets, briefly summarized in Tab.1, were chosen due to their face images contain variations in illumination, expressions,postures, and occlusions in which we aim to tackle these problems.All experiments were executed using Python programming language on a personal computer with the following specifications:Intel(R) Core (TM) i7-8550u CPU @ 1.80 GHz 1.99 GHz with 12.00 GB RAM.

Table 1:Brief description of ORL, Yale, GT, and FEI face datasets

Before conducting the experiments, face images of the four used face datasets were converted to grayscale and then resized to 64×64 pixels.Only face images of the Yale face dataset were manually cropped prior to resizing.To investigate the efficiency of the proposed method under different numbers of training samples per person, we conducted six independent experiments on each face dataset.Where, we randomly selected N-facial images (N=4, 5, 6, 7, 8, 9) for each person as training samples, and then we used the remaining facial images as testing samples.To recognize the unseen test face images, the nearest distance classifier is employed with one of the six well-known distance measures found in the literature, namely Bray-Curtis, Euclidean, Manhattan,Mahalanobis, Correlation, and Cosine.Each of these distance classifiers was used independently with the proposed method to see which of them can give a higher recognition rate.

To measure the performance of the proposed method, we used cumulative match characteristic (CMC) curves [15,39].CMC curve is employed to judge the ranking capabilities of the proposed method.Due to the random selection of training samples, the proposed method obtains different recognition rate for each run.Therefore, to avoid biases, the proposed method, for each experiment, takes the mean of ten independent runs as the average rank-one recognition rate.For FastICA and LDA methods, we used built-in functions of the Python programming language with its default parameters.We only changed the number of the components for the FastICA method to be 150, 115, 75, and 40 components across the four face datasets FEI, GT, ORL, and Yale, respectively.This is because the updated configuration has given us the best results based on trial-and-error tests.In the following subsections, first, the experiments on each of these four face datasets used are highlighted.Then, the results of the comparisons with recent face recognition approaches are presented and discussed.

5.1 Experiments on GT Face Dataset

Georgia Tech (GT) face dataset [40] includes face images of fifty persons, with fifteen different face images per person.This dataset was generated at the Georgia Institute of Technology in two or three sessions at different time intervals.The captured face images of this dataset vary in lighting, facial expression, size, and rotation.Fig.5 shows some facial image samples of three individuals from this dataset.

Figure 5:Some samples of face images from the GT face dataset

Fig.6 presents the average rank-one recognition rates of the six distance measures used.Choosing N=5 randomly, the average CMC for the six distance measures used are also shown in Fig.7.It can be seen from Fig.6 that the highest rank-one recognition rates for N-training samples (N=4,5,6,7,8,9) per person are 81.38%, 86.42%, 89.20%, 91.37%, 92.80%, and 93.50%,respectively.For instance, when selecting N=4 randomly, the highest rank-one recognition rate is obtained by the Bray-Curtis distance classifier, which is 81.38%.From the results in Figs.6 and 7, it can be noticed that Bray-Curtis and Cosine distance classifiers perform better than Correlation, Euclidean, Manhattan, and Mahalanobis distance classifiers as they yield superior results.However, Cosine distance classifier gives better results than all other distance classifiers used when selecting N= 6,7,8,9 randomly (see Fig.6).On the contrary, Bray-Curtis distance classifier achieves the highest rank-one recognition rates when selecting N=4,5 randomly.

Figure 6:Rank-one recognition rate on GT face dataset

Figure 7:CMC curves of six distance measures on GT face dataset

5.2 Experiments on ORL Face Dataset

ORL face dataset [41] contains face images, in PGM format, for each of 40 distinct individuals, with ten different face images for each individual.The captured face images of this dataset vary in illumination, facial expressions (closed/open eyes, not smiling/smiling), and facial details(no glasses/glasses).Fig.8 shows some facial image samples of three individuals from this dataset.

Fig.9 shows the average rank-one recognition rates of the six distance measures used.Choosing N=5 randomly, the average CMC for the six distance measures used are also depicted in Fig.10.From Figs.9 and 10, it can be observed that the Bray-Curtis distance classifier attains better results than all other distance classifiers used when the number of training samples is either 4, 5, 6, 7, or 9 samples per person.Conversely, Correlation and Cosine distance classifiers achieve the highest rank-one recognition rates when the number of training samples is eight samples per person (see Fig.9).This shows that Bray-Curtis, Correlation, and Cosine distance classifiers perform better than Euclidean, Manhattan, and Mahalanobis distance classifiers as they provide superior results.

Figure 8:Some samples of face images from the ORL face dataset

Figure 9:Rank-one recognition rate on ORL face dataset

Figure 10:CMC curves of six distance measures on ORL face dataset

5.3 Experiments on Yale Face Dataset

Yale face dataset [42] comprises of different face images, in GIF format, for each of fifteen distinct subjects, with eleven different face images for each subject.The captured face images of this dataset also vary in illumination and facial expressions (normal, happy, left-light, center-light,wearing glasses, wearing no glasses, right-light, sad, surprised, wink, and sleepy).Fig.11 shows some facial image samples of three individuals from this dataset.

Figure 11:Some samples of face images from the Yale face dataset

Fig.12 reports the average rank-one recognition rates of the six distance measures used.Choosing N=5 randomly, the average CMC for the six distance measures used are also described in Fig.13.It is evident from the results in Figs.12 and 13 that the Mahalanobis distance classifier outperforms all other distance classifiers used when selecting N=4, 5 randomly.Nevertheless,Cosine distance classifier achieves the highest rank-one recognition rates when selecting N=6,7, 9 randomly (see Fig.12).On the other hand, Correlation distance classifier achieves the highest rank-one recognition rates when selecting N = 7, 8 randomly.This indicates that the number of training samples determines the choice of the distance classifier which gives the best recognition rate.

Figure 12:Rank-one recognition rate on Yale face dataset

Figure 13:CMC curves of six distance measures on Yale face dataset

5.4 Experiments on FEI Face Datasets

FEI face database [43] comprises of different face images for each of 200 distinct subjects,with 14 different face images for each subject.All face images are captured against a white homogenous background in an upright frontal position with profile rotation of up to about 180 degrees.Fig.14 shows some facial image samples of three individuals from this dataset.

Figure 14:Some samples of face images from the FEI face dataset

Fig.15 outlines the average rank-one recognition rates of the six distance measures used.Choosing N=5 randomly, the average CMC for the six distance measures used are also reported in Fig.16.It is observed from Figs.15 and 16 that the Cosine distance classifier yields superior results compared to all other distance classifiers used regardless of the number of training samples chosen for each person.On the other hand, Mahalanobis distance classifier performs worse than other distance classifiers used as it provides the lowest rank-one recognition rates.

Figure 15:Rank-one recognition rate on FEI face dataset

5.5 Discussion

From Figs.6, 9, 12, and 15, it can be seen that in almost all cases, the recognition rate increases with the increase in the number of training samples regardless of the distance classifier used.This is reasonably acceptable because of the high probability of predicting the precise class.It can also be observed from the results in Figs.6, 9, 12, and 15 that the recognition rate varies from one distance classifier to another.This indicates that the proposed method is consistent and sensitive to the distance measures used.

The results obtained from ORL, Yale, FEI, and GT face datasets confirmed that the proposed method is efficient under conditions in which facial images vary in illumination, expressions,postures, and occlusions, as these face datasets contain the above challenges.Moreover, choosing a distance classifier that gives the best recognition rate depends on the number of training samples used.

Figure 16:CMC curves of six distance measures on FEI face dataset

5.6 Comparison with Other Approaches

In this section, the results of the proposed method are compared with the recent approaches in the literature.To make the comparisons fair, we used the same number of training samples as those used in the recent methods.Besides, we chose the results of one distance classifier with the proposed method to be compared with other approaches across the four face datasets:Yale,ORL, FEI and GT.The chosen distance classifier is Cosine due to its high recognition rate among almost all the face datasets used.Tabs.2-5 show the comparison results of the proposed method with the recent approaches on ORL, Yale, GT, and FEI face datasets, respectively.

It is evident from the results in Tab.2 that the proposed method has obtained the highest recognition rate compared to the approaches presented in [13-17,20-22] across ORL face dataset.For instance, when the number of training samples is five, our proposed method provides a recognition rate of 99.10% which is significantly higher than those provided by other methods:97.00% by [15], 94.50% and 97.85% by [17], 97.75% by [21], and 97.50% and 95.50% by [22].The improvement noticed in the recognition rate by our new method is about 1.25, 2.19, and 1.13 points compared to the highest recognition rate provided by other methods across the three N-training samples (N=5,6,8), respectively.This indicates that the proposed method is efficient and provides better results than the methods presented in [13-17,20-22] for the ORL face dataset.

Table 2:Comparison of the proposed method with state-of-the-art approaches on ORL face dataset

Table 3:Comparison of the proposed method with state-of-the-art approaches on Yale face dataset

From Tab.3, it can be seen that the proposed method has attained the highest recognition rate compared to the methods presented in [13,14,17,20,22] across Yale face dataset.The improvement in the recognition rate by our new method is about 0.12 and 0.84 points compared to the highest recognition rate reported by other methods across the two N-training samples (N=5,8),respectively.This shows that the proposed method is also effective and yields superior results than the methods presented in [13,14,17,20,22] for the Yale face dataset.

It is noticeable from the results in Tab.4 that the proposed method has achieved the highest recognition rate compared to the methods presented in [15,16,19,21,23].The improvement in the recognition rate by our new method is about 9.48, 13.08, 11.17, and 5.2 points compared to the highest recognition rate given by other methods across the four N-training samples (N=7,8,9,10),respectively.This demonstrates that the proposed method is also effective and provides better results than the methods presented in [15,16,19,21,23] for the GT face dataset.

Table 4:Comparison of the proposed method with state-of-the-art approaches on GT face dataset

From Tab.5, it can be noted that the proposed method has attained the highest recognition rate compared to the methods presented in [13-15,18,24] across FEI face dataset.The improvement in the recognition rate by our new method is about 4.17, 0.2, and 1.21 points compared to the highest recognition rate reported by other methods across the three N-training samples(N=7,8,11), respectively.This demonstrates that the proposed method is also effective and yields superior results than the methods presented in [13-15,18,24] for the FEI face dataset.

Table 5:Comparison of the proposed method with state-of-the-art approaches on FEI face dataset

Based on the above experimental results, we can conclude that the proposed method outperforms state-of-the-art methods presented in the literature across the four benchmarked face datasets:ORL, Yale, FEI, and GT investigated in this paper.The main reason behind this improvement in our approach is the combination of the three well-known methods of feature extraction together to extract the most informative features.Another reason for this improvement is the use of the Cosine distance classifier to calculate the recognition rate in which it has shown to give a higher recognition rate than the other most known distance classifiers.

6 Conclusions and Future Works

This paper proposes a new approach to face recognition based on three feature extraction methods, namely Gabor-based feature extraction, FastICA, and LDA.The proposed method was assessed on four benchmarked face datasets:GT, ORL, FEI, and Yale.Six distance measures(e.g., Euclidean, Cosine, Bray-Curtis, Mahalanobis, Correlation, and Manhattan) were employed to investigate the performance of the proposed method.According to the experimental results, the proposed method achieves the highest rank-one recognition rate when compared to state-of-theart methods in the literature.This indicates that our fusion of Gabor filters with FastICA and LDA gives a better result compared to other fusion methods suggested in the literature in terms of recognition rate.Moreover, the results of our experiments illustrated the efficacy of the new proposed method, especially under using Cosine distance measure.

The method presented in this paper appears to be robust and sensitive in identifying facial images under conditions where lightings, facial expressions, occlusions, and rotations vary, as it provides superior results for ORL, Yale, and GT, FEI face datasets.However, the main drawback of our proposed method is that it is time consuming and computationally expensive due to the use of three sophisticated feature extraction methods.To overcome this issue, research will be further extended in the future to investigate other possible fusion methods, such as combining Gaborbased feature extraction with recent bio-inspired feature selection methods like the Cuttlefish-based method presented in [44].Moreover, further investigation can be performed on facial images captured in more noisy environments with more variations in illuminations, facial expressions,occlusions, and rotations.

Funding Statement:The authors received no specific funding for this study.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

主站蜘蛛池模板: 99在线视频网站| 在线无码九区| 99青青青精品视频在线| 日本道综合一本久久久88| 无码免费的亚洲视频| Jizz国产色系免费| 欧美不卡在线视频| 久久精品中文字幕免费| 国产毛片久久国产| 免费a在线观看播放| 九九热精品视频在线| 精品偷拍一区二区| 日本久久网站| 一本无码在线观看| 久久一本日韩精品中文字幕屁孩| 天堂成人在线| 丰满人妻被猛烈进入无码| 国产v精品成人免费视频71pao | 国产欧美日韩另类| 69av免费视频| 亚洲综合18p| 免费99精品国产自在现线| 毛片免费视频| 国产白浆一区二区三区视频在线| 成年A级毛片| 囯产av无码片毛片一级| 国产在线专区| 高清大学生毛片一级| 成人a免费α片在线视频网站| 免费a级毛片视频| 中文字幕无线码一区| jizz在线免费播放| 日本精品视频一区二区 | 欧美乱妇高清无乱码免费| 尤物午夜福利视频| a在线亚洲男人的天堂试看| 国产区91| 色久综合在线| 在线毛片免费| 亚洲成年网站在线观看| 亚洲av无码成人专区| 九色综合伊人久久富二代| 久久久久亚洲精品成人网| 无码AV动漫| 欧美福利在线观看| 黑人巨大精品欧美一区二区区| 欧美伦理一区| 波多野吉衣一区二区三区av| 欧美福利在线| 精品无码国产自产野外拍在线| 国产色婷婷| 一区二区三区国产精品视频| 欧美精品黑人粗大| 高潮毛片无遮挡高清视频播放| 少妇高潮惨叫久久久久久| 亚洲精品大秀视频| 国产日韩欧美中文| 欧美特级AAAAAA视频免费观看| 在线观看国产一区二区三区99| 久久女人网| 看国产毛片| 欧美日本中文| 国产农村1级毛片| 欧美成人一级| 久久一色本道亚洲| 超碰色了色| 国产一区二区免费播放| 福利视频一区| 久久精品人人做人人综合试看 | 国产精品视频猛进猛出| 91在线国内在线播放老师| 青草视频网站在线观看| 韩日免费小视频| 国产精品入口麻豆| 欧洲精品视频在线观看| 欧美视频在线第一页| 国产手机在线小视频免费观看 | 国产精品嫩草影院av| 国产欧美亚洲精品第3页在线| 国产精品美乳| 成人国产免费| 好紧太爽了视频免费无码|