Inter-Communication Classification for Multi

Transcription

Inter-Communication Classification for Multi
The International Arab Journal of Information Technology, Vol. 11, No. 4, July 2014
387
Inter-Communication Classification for MultiView Face Recognition
Chouaib Moujahdi1, Sanaa Ghouzali1, 2, Mounia Mikram1, 3, Abdul Wadood4, and Mohammed Rziza1
1
LRIT (Associated Unit with the CNRST), Mohammed V-Agdal University, Morocco
2
Information Technology Department, King Saud University, Saudi Arabia
3
The School of Information Sciences, Mohammed V-Agdal University, Morocco
4
Computer Engineering Department, King Saud University, Saudi Arabia
Abstract: In this paper we present a new multi-view face recognition approach. Besides the recognition performance gain
and the computation time reduction, our main objective is to deal with the variability of the face pose (multi-view) in the same
class (identity). Several new methods were applied on face images to calculate our biometric templates. The Laplacian
Smoothing Transform (LST) and Discriminant Analysis via Support Vectors (SVDA) have been used for the feature extraction
and selection. For the classification, we have developed a new inter-communication technique using a model for the automatic
pose estimation of the head in a face image. Experimental results conducted on UMIST database show that an average
improvement for face recognition performance has been obtained in comparison with several multi-view face recognition
techniques in the literature. Moreover, the system maintains a very acceptable running time and a high performance even in
uncontrolled conditions.
Keywords: Face recognition, multi-view, inter-communication, LST, SVDA, pose estimation.
Received July 26, 2012; accepted March 19, 2013; published online April 4, 2013
1. Introduction
Facial recognition is a biometric technique based on
physiological/behavioral characteristics specific to each
person. These characteristics have the advantage that
they are universal, unique, permanent and cannot be
falsified, unlike conventional techniques such as
passwords and badges that can be used fraudulently by
others. For the human brain, the process of face
recognition is a high-level visual task. Although
humans can detect and identify faces in a scene without
much trouble, to build an automatic system that
performs such tasks represents a serious challenge. This
challenge becomes greater when the conditions of
image acquisition are highly variable and
uncontrollable. There are two types of variations
associated with face images: inter- and intra-subject.
The inter-subject variation is limited because of the
physical
resemblance
between
individuals.
Contrariwise, the intra-subject variation is larger and it
can be attributed to several factors: Variations in facial
expressions, lighting conditions, occlusions, unwanted
noise, affine distortions and clutter, etc, may give some
bad impact on the overall performance of face
recognition accuracy. But in general, these problems
are solved with a very high percentage [4, 9, 10, 24].
However, multi-view face recognition, as shown in
Figure 1, remains a major challenge for building a
robust/ reliable biometric system, able to achieve high
recognition rate. This problem has attracted the efforts
of researchers in recent years and is the objective of
our research as well. This paper presents some
development results of a new multi-view face
recognition approach. In this work, several new
approaches are used. A very useful approach to build
our proposed technique of inter-communication
between classifiers is a model for pose estimation to
specify the angle of view in a face image [1]. The
other is the Laplacian Smoothing Transform [12],
used to extract low frequency features, followed by
Discriminant Analysis via Support Vectors to reduce
the dimensionality of the extracted feature vectors
[13].
The rest of the paper is organized as follows:
Section 2 presents an overview of multi-view
approaches. Section 3 describes the techniques of
feature selection LST and SVDA and the proposed
approach. Experimental results are discussed in
section 4, conclusions and perspectives are drawn in
section 5.
2. Overview of Multi-View Approaches
Multi-view face recognition remains a major
challenge to build a robust recognition and a reliable
identification. This problem is a subject of active
research in the fields of pattern recognition and
computer vision. It has attracted research efforts both
because of its potential applications and the challenge
it presents. To ensure robust multi-view face
388
The International Arab Journal of Information Technology, Vol. 11, No. 4, July 2014
recognition with high recognition rate, some
remarkable approaches are proposed in [7, 8, 15, 19].
According to the used technique, we can divide these
approaches into three main categories: Feature-based
approaches, classification-based approaches and hybrid
approaches.
Figure 1. Example of the pose variation from UMIST database.
2.1. Feature-Based Approaches
We can distinguish two type of feature-based
approaches: linear and non-linear approaches. The
linear approaches are based on methods that rely on
statistical properties and use linear algebra. [19]
presents a work based on a linear approach where a
new multi-view database is created, using a simple
acquisition system which consists of five cameras able
to capture simultaneously five views of a human face
with different observation angles. In this work, feature
selection was applied in the spatial domain using
traditional linear/statistical techniques: Principal
Component Analysis (PCA) [29] followed by
Independent Component Analysis (ICA) [2] on the
training and test images. The results shown in this study
prove that these methods are ineffective for non-linear
problems. Some studies have shown that the human
visual system is more sensitive to variations in the lowfrequency band, and that face recognition performance
can be increased in the frequency domain [6]. This has
led researchers to work in the frequency domain. In
[15], a multi-view face recognition based on Gabor
wavelet transform is presented. The face images of
multiple views are convolved with a bank of Gabor
wavelet filters. These convolutions produce high
dimensional feature vectors (Gabor faces). The
dimensionality reduction of Gabor faces is achieved by
Linear Discriminant Analysis (LDA) [16]. Unlike PCA
algorithm, LDA performs a true separation of classes to
reduce the space of facial features in a low dimensional
space. However, the use of LDA, which has several
problems and weaknesses (see subsection 3.1.2.), can
degrade the overall multi-view system performance.
Multi-view face recognition is a non-linear multiclass problem, several studies have shown that linear
methods are inefficient for such problems. In recent
years, other researchers have tried to develop
techniques like Kernel-PCA (KPCA) [27] and Kernel
Direct Discriminant Analysis (KDDA) [20] to make the
distribution patterns of faces simple and linear,
consequently, making them applicable to non-linear
problems. These techniques combine the strengths of
traditional linear methods and kernel machines to
solve non-linear problems. In [21], KDDA with an
RBF kernel (Radial Basis Function as kernel) was
used for feature selection. However, these kernel
methods can not always effectively address the Small
Sample Size (SSS) problem [17] and their
classification process is very slow because of the
computations in the kernel machine. Thus, their results
are still modest. In [7], a new non-linear method is
proposed: Support Vector Machines-Discriminative
Feature Selection (SVM-DFS), to select most
discriminative features without linearly combining the
original features. SVM-DFS proved its robustness
compared to KPCA and KDDA. This approach
reduces the computational time using a linear kernel
but with a degradation of recognition rate and it keeps
a higher recognition rate by using an RBF kernel
while increasing the test time. Thus, it is clear that
SVM-DFS cannot find a better compromise between
performance and computation time.
2.2. Classification-Based Approaches
Feature based approaches in the previous section
focused on the extraction of significant features
whereas the classification process has been conducted
using either K-Nearest Neighbors (KNN) or Support
Vector Machines (SVM) classifiers. Other researchers
have focused on the development of new classification
methods to solve the problem of multi-view face
recognition, which it is a multi-class classification
problem. The SVM has been successfully applied to
various classification problems, but several large-scale
problems are overly difficult to solve using traditional
SVMs. In [8], multi-class SVMs [26] was used to
produce a new method: Min-Max Modular SVMs
(M3-SVM) for multi-view face recognition. In general,
the working procedure of M3-SVM consists of three
stages. The complicated problem of multi-view face
recognition was decomposed into several relatively
simpler two-class sub-problems, and each subproblem is transformed into a binary problem using
the one-versus-one strategy (tasks decomposition
step). For each sub-problem an SVM based
Discriminative Feature Selection (SVM-DFS) method
[7] was used (sub-classification step). Finally, the
combination of sub-solutions strategy follows the
minimization and the maximization principles [22]
(solutions reassembly step). The results obtained in
this study are still modest.
2.3. Hybrid Approaches
Some works in multi-view face recognition make use
of more than one basic approach using together
feature techniques (linear and non-linear) and
Inter-Communication Classification for Multi-View Face Recognition
classification techniques simultaneously. We refer to
such techniques as hybrid approaches. The objective of
the work [28] is to create a new generative model,
which can handle both linearity (identity) and nonlinearity (view) of face images in multiple views,
combining manifold learning [18] and tensor analysis.
The major problem with this approach is that the
treatment of the non-linearity of the view manifolds can
influence the identity information, which can degrade
the overall performance of the system. To solve this
problem, an algorithm to calculate a heuristic parameter
useful for facial recognition was proposed in [11]. The
experiments were performed on Weizmann and oriental
face database. This approach has proved its robustness
over the traditional tensor-based methods like
TensorFace [31] and VPCA [25], but it did not improve
upon the results of other approaches [7, 15].
3. Proposed Approach
In this section, we present a hybrid approach of multiview face recognition based on new methods of feature
selection and a technique of inter-communication of
classifiers using a model of pose estimation in 2D
images.
3.1. Feature Selection
To build robust learning models, feature selection is
used to extract the most relevant features and remove
the most redundant. Motivated by previous studies [6],
we have decided to work in the frequency domain. For
feature selection, we applied the efficient feature
extraction method LST [12] followed by the multi-class
dimensionality reduction approach SVDA [13].
3.1.1. Laplacian Smoothing Transform
Gu et al. showed in [12] some drawbacks of statistical/
manifolds [33, 34] methods resulting due to the
limitation of each pixel being considered independent
and the links with its neighbors are not considered. To
cope with these limitations, Laplacian Smoothing
Transform (LST) is used to represent the image in the
frequency domain. Thus, the high frequency of the
image (containing noise) can be eliminated while
keeping critical data that are represented by low
frequencies and keeping the connection between the
pixels. Compared to Discrete Cosine Transform (DCT)
[32] and Discrete Wavelet Transform (DWT) [5]
methods with respect to efficiency and speed, LST has
proven its robustness/priority [12]. For the LST, the
computational complexity lies in the number of
eigenvectors of the Laplacian matrix L (Expression
3) to be calculated for images of size M×N. However,
the Laplacian matrix L which is a sparse matrix can
be calculated quickly and once for training. Generally,
the LST can be an effective method for feature
extraction in face recognition. The stages of
development of LST are:
389
• Calculate the weight matrix W of size MN× MN (M
and N are the dimensions of the image):
1 if x − x' + y − y' = 1

W ( x , y , x' , y' ) = 
'
'
1 if x − x + y − y ≠ 1
Where x , y = x × N + y
(1)
x and y are the coordinates of a pixel.
• Calculate D, a diagonal matrix whose entries are
column (or row, since W is symmetric) sums of W:
D( x, y, x' , y' ) = ∑W ( x, y, x' , y' )
x, y
(2)
• Calculate the Laplacian Matrix:
LMN = D − W
(3)
• Compute first k eigenvalues and the k
corresponding eigenvectors Ek (k depends on the
number of adequate low frequencies).
• Project the training matrix F (which contains
training images: one image in each row) in the
space spanned by the eigenvectors Ek to build the
training set of signatures G after pre-processing
with LST:
G = E kT F
(4)
• Similarly, each test image f is proposed to construct
the signature test g after preprocessing by LST:
g = E kT f
(5)
3.1.2. Discriminant Analysis via Support Vectors
Many studies have shown that the statistical methods,
such as LDA, in the frequency domain would improve
the recognition rates [34]. However, LDA has several
problems, for example it suffers from the problem of
Small Sample Size (SSS) and it creates subspaces that
favor classes well separated from those which are not.
Thus LDA fails to obtain the optimum direction for
separating two classes in all situations. Alternately, the
SVM can discover the optimal directions to maximize
the margin between two classes [30]. Hence, the idea
to combine LDA with SVM to build a robust multiclass approach for dimensionality reduction:
Discriminant Analysis via Support Vectors (SVDA).
First, SVM is used to calculate an optimal direction to
discriminate between two classes, then the criteria of
class separability (similar to LDA) are calculated
using the distinct Support Vectors (SV). Finally, the
projection matrix is calculated. The general stages of
SVDA development are:
• For every two classes cl and cm, 1 ≤ l < m ≤ M
(number of classes), an SVM is employed to find
an optimal direction clm. At the end of this
procedure, we will have a matrix Φ, which contains
M ( M − 1 ) optimal directions clm (a direction by
2
column).
390
The International Arab Journal of Information Technology, Vol. 11, No. 4, July 2014
• The between-class matrix Vb is given by:
Vb =
∑
(6)
T
c lm c lm
= ΦΦ T
1 ≤i < j ≤ M
• Similar to LDA, the within-class matrix Vw is given
by:
M
(7)
V =
( x̂ − µˆ )( x̂ − µˆ )T
w
∑∑
m =1 i∈Î m
i
m
i
m
x̂ i : The vectors of the data matrix of SVs.
µ̂ m : The average of SVs in the class cm.
Î m : Indices of Support Vectors in the class cm.
• Searching for the optimal projection by solving the
equation:
V b α = βV w α
(8)
α: The set of eigenvectors.
β : The set of eigenvalues.
• Eigenvectors corresponding to the largest k'
eigenvalues form the columns of the final
transformation matrix.
In summarize, we presented the techniques used for
feature selection: LST followed by SVDA. For
classification, we propose the technique of intercommunication classification. To apply this technique
we need a face pose estimator. In the next axis, we
present the estimation model used in our work.
3.2. Pose Estimation
The automatic pose estimation of faces in a 2D image is
usually done in two stages: first detect the face and then
the pose of the face is estimated. There are several
different approaches to solve this problem, which can
be divided into three categories [3]: Geometric methods
[23], genetic methods [14], and learning methods [1].
The first category requires the use of 3D information
provided by the image sensor to estimate the pose, but
it is difficult for our work and all available multi-view
databases do not provide this information. In the second
category, the pose will not be calculated in real time
given the high complexity of genetic algorithms.
Therefore, this class is not suited to our needs. For the
latter category, it uses the technique of artificial
learning of 2D images. In this case, a sufficiently large
set of face pictures (with different poses from -90° to
+90°) must be presented to the system during a learning
phase. Once the learning phase is complete, the system
should be able to estimate the pose of a given test
image. Generally, this category has a running time quite
acceptable.
In this work we used the pose estimation model
given by [1]. This model is a learning-based method
which treats the problem of pose estimation as a
regression problem and not as a classification problem
by assigning the face to a class of many poses. In
addition, this model works on images of wide
variations in the background, lighting and expression.
We can summarize the principle and the general
operation of the model in four steps:
• The test image (of size 60 × 60) is divided into a
grid of blocks of size 10 × 10 (i.e., divided into 36
blocks).
• A library of facial images (240 images), with a
range of poses from -90° to +90°, is used. The
library can be regarded as a palette from which the
image blocks can be taken. These blocks provide
information about the true pose.
• The library is used to approximate each block of the
test image. Finally, an approximated image is
formed.
• Pose parameters W are used to interpret the
selected blocks using the Bayesian posterior
probability:
Pr ( β | Y , W ) =
∏iP=1 Pr ( yi , l* | β , ωi ) Pr ( β )
Pr ( Y )
(9)
β: The pose and Y is the test image, Y=[y1, …, yp]
(decomposed into P blocks).
l*: The block of the library that is closest to the test
block yi.
ωi: Denotes the vector of parameters associated with
the ith: Block of the test image and all the blocks in
the library. For more details about the construction of
W you can consult [1].
One of the drawbacks of this model is that it cannot
estimate view angles, greater than 70° and lower than 70°, with high accuracy. However, in practice, this
estimation is sufficient for our needs. Our proposed
inter-communication technique (especially the angles
quantization strategy) makes the incorrect estimations
bearable.
3.3. Inter-Communication Classification
In multi-view face recognition as shown in Figure 2,
sometimes the distance Y between the images (with
different views) of a same identity is greater than the
distance X between the images of the same class of
view (with different identities), which requires
treatment of the variability of the view in the same
class (identity). It is the purpose of our proposed
technique.
After building the training and the test databases,
we apply the LST followed by the SVDA for the
feature selection to extract comparison signatures
associated with each image. For classification, we
implement an inter-communication technique using
the model of pose estimation developed by
Aghajanian and Prince [1]. Since the angles are
estimated like real values in a very wide range, we
have applied a quantification technique of the results,
as shown in Figure 3, which has led to six classes of
Inter-Communication Classification for Multi-View Face Recognition
θ1 =θ-0°= 20° and θ2 =θ-15°= 5°
θ2 =θ-30°= 20° and θ4 =θ-45°= 25°
θ5 =θ-60°= 40° and θ6 =θ-75°= 55°
angles (0°, 15°, 30°, 45°, 60°, and 75°). The test
signature is compared to the training signatures already
prepared using six KNN classifiers. Each classifier has
a class of angles responsible for comparing the test
signature with the training signatures that belong to its
class and returns a distance. We use a classification
with a weighted majority vote on the six obtained
distance values.
45°
• In this case, the minimum distance is reached for
the classifier 15°. The order of priority of classifiers
is as follows: Classifier 15°, classifier 30°,
classifier 0°, classifier 45°, classifier 60° and
classifier 75°.
• The weight of each classifier is calculated using the
following expression:
30°
60°
15°
Identity 1
75°
391
0°
Pi = 1 −
X<Y
X
for i ∈ [ 1, 6 ]
(10)
j
• Finally, we organize a majority vote. For classifiers
that return the same subject class, we sum its
weight. The largest vote favors its identity (subject
class) as a final result.
Identity 2
Figure 2. The View variations of two different identities and
comparison between two distance types (X and Y) in multi-view
face recognition.
It is to be noted that the proposed approach is valid
with any type of classifier. We chose to use the KNN
classifier as it is characterized by simplicity and speed
which are very important for our work. On the other
hand, Gu proved in [13] that SVDA is adapted with
KNN and can improve widely its classification task.
Pose estimation: -19
Pose
estimation
Angle θ
Image
4. Experimental Results
if
θ = 75°
∑θ
i =1
Y
θ > 67.5°
θi
6
θ є ]52.5,67.5]
θ = 60°
θ є ]37.5,52.5]
θ є ]22.5,37.5]
θ = 45°
θ = 30°
θ є ]7.5,22.5]
θ = 15°
In this Section, the experimental results of the
proposed method on UMIST face database are
described. UMIST face database consists of 550 face
images of 20 distinct persons/ subjects. Faces in the
database cover a wide range of poses from profile
(90°) to frontal (~0°) view. The database covers also a
mixed range of race, gender and appearance, such as
different expressions, illuminations, glasses/no
glasses, beard/no beard, different hair style, etc.
To prove the efficiency of inter-communication of
classifiers, we compare the proposed system with a
traditional system using LST (69 low frequency)
θ ˂ 7.5°
θ = 0°
Figure 3. The strategy for quantification of image angles.
Figure 4 summarizes the general operation of the
approach. Suppose for example that the angle in a test
image is θ = 20°:
• We calculate the values θi to specify the priority of
classifiers, as follows:
Pose estimation: 20
Pose
estimation
Angle θ
Pose
estimation
Test image
Test
signature
Quantification
(Figure 3)
Training database
LST + SVDA
Training
signatures
Six weighted KNN
classifiers
(expression 10)
75°
60°
0°
15°
.
.
.
.
.
.
.
.
.
.
.
.
45°
30°
75°
0°
Quantified angles
majority vote
Final result
Figure 4. The general operation of proposed approach.
392
The International Arab Journal of Information Technology, Vol. 11, No. 4, July 2014
Error rate(%)
followed by SVDA and a single KNN classifier. Test
conditions are not ideal: each person is presented with
six images (one image per view class) and the
particular case of cross-validation, leave-one-out, is
used as a testing technique. In leave-one-out, the
algorithms run N times. In each round, N-1 samples are
used for training and the remaining sample is used for
testing. If the tested sample is correctly predicted, the
test accuracy of the round is 100%, otherwise it is 0%.
The test accuracy of the leave-one-out strategy is the
mean accuracy of all the N predictions. Which means
that in our case, in each round, the view class of the test
image is not present in the training database, which
presents a complex test situation.
As shown in Figure 5, between 1 and 4 SVDA
eigenvectors the results are similar, it is evident that the
information is still insufficient. Between 4 and 10
eigenvectors we can notice a small improvement in our
system, and for 10 eigenvectors and more the efficiency
of our classification technique is clear. We found that
the inter-communication between classifiers increases
the robustness of the system by 4.17%.
identified the true person, so after the majority vote
they won the vote with a weight of 2.6567 (0.97015 +
0.95522 + 0.73134) versus the other classifiers, 45°,
30°, and 0°, that have returned the same wrong
identity with a weight of 2.34329 (0.8806 + 0.80597 +
0.65672). Therefore, the person is correctly identified.
Total weights: 2.6567
Sought identity
Test image
Pose estimation: 69
Aligned image
a) Pose estimation and the final result after the majority vote.
Weight: 0.97015
Classifier 75°
Weight: 0.95522
Classifier 60°
Weight: 0.8806
Classifier 45°
Weight: 0.80597
Classifier 30°
Weight: 0.73134
Classifier 15°
Weight: 0.65672
Classifier 0°
b) Result of each classifier and the weights used in the majority vote.
b) Result of each classifier and the weights used in the majority vote.
Figure 7. Test using the proposed system.
SVDA Egenvectors
Figure 5. Comparison between the proposed system and a
traditional system.
Figure 6 shows a sample test with a profile view
using the traditional system (without intercommunication of classifiers). A single classifier is
used for the classification, LST (69 low frequency)
followed by SVDA (14 eigenvectors) are used for
feature selection, and test conditions are non-ideal. This
system did not give the correct results.
Test image
Results image
Wrong result
True result
Figure 6. Test using traditional system.
Figure 7 shows the operation of an authentication by
our system using the same profile view in Figure 6.
Classifier 75° has the highest priority because it is the
closest to the estimated angle which equals 69°,
therefore it has the largest weight: 0.97015 (calculated
using Expression 10). Classifiers 75°, 60° and 15° have
It is to be noted that it is possible to use only the
classifiers close to the view class, to minimize the
complexity. But this trick will work only if the
conditions are ideal and it can cause considerable
degradation of system performance when used for
example in the test conditions similar to those
presented in Figure 5. Our experience has shown that
the use of six classifiers in the majority vote is needed,
especially in non-ideal test conditions.
To validate the experimental results obtained by the
proposed system, we have presented in Table 1 a
comparison with several works on multi-view face
recognition using similar test conditions. We divided
UMIST database into two subsets: the training set of
200 images, such as 10 images per person are
carefully selected according to the head pose, and the
test set contains the remaining 350 images. The size of
each cropped image in all the experiments is 92×112
pixels. Thus, each image can be represented by a
10304-dimensional vector in image space. Each face
image vector was normalized to unit before use (to
have unit variance). We reduced the input
dimensionality (originally at 10304) by projecting the
data onto its 69 LST low frequency coefficients. The
final transformation matrix contains 14 SVDA
eigenvectors corresponding to the largest 14
eigenvalues.
Inter-Communication Classification for Multi-View Face Recognition
In these test conditions, we have achieved a
recognition rate of 100%. Former works on multi-view
face recognition have been occupied either by
developing feature selection techniques to minimize
error rate or by developing classification techniques to
minimize the runtime. Consequently, the balance
between recognition rate and running time was never
achieved in these works. Contrariwise, results of
Table1 show an increase in recognition performance
with our system over previous works, keeping a very
acceptable runtime.
Table 1. Comparison between our approach and some other works.
Criteria
Test Base
Recognition
Rate
Runtime
KDDA [21]
UMIST
94.3%
---
SVM-RBF [15]
UMIST
96.7%
---
UMIST
96.7%
(Linear kernel)
98.2%
(RBF kernel)
Technique
SVM-DFS [7]
463 ms
4.16 s
M -SVM [8]
UMIST
93.1%
1.65 s
Proposed
UMIST
100%
3.32 s
It should to be noted that incorrect pose estimation
may influence the recognition performance of the
proposed system, because the weights of classifiers will
be miscalculated. Therefore, the majority vote may
promote a wrong result. Incorrect pose estimation takes
place, mainly, if the detection of face is inaccurate. In
the case of UMIST database, all face images are
cropped and well detected, which means that all pose
estimations are relatively correct.
5. Conclusions
In this paper, we proposed a new system of multi-view
face recognition. For the features selection, we
proposed the Laplacian Smoothing Transform (LST) to
extract the low frequency smooth features of an image
and high frequencies (containing noise) can be removed
while keeping the dependency between neighboring
pixels. We proposed to follow LST by Discriminant
Analysis via Support Vectors (SVDA), the new multiclass approach, to reduce the dimensionality. We used
information provided by the model of the automatic
pose estimation to propose a technique of intercommunication between multiple KNN classifiers, to
improve the efficiency and robustness of the system.
This work can be considered as a stepping stone for
future work in this research direction. The future works
are suggested to optimize the pose estimation model to
minimize the execution time and use other larger
databases to validate the performance of the proposed
system. We are also planning to propose and test
various other techniques of weighted majority vote and
make a non-linear SVDA. Then, for further work, we
will focus on other areas of research related to an
important complement of biometrics: Biometric
security protection.
393
Acknowledgements
This work was supported by the Research Center of
College of Computer and Information Sciences, King
Saud University. The authors are grateful for this
support.
References
[1]
Aghajanian J. and Prince S., “Face Pose
Estimation in Uncontrolled Environments,”
International Conference on Computer Vision,
available at: http://www.bmva.org/bmvc/2009/
Papers/Paper256/Abstract256.pdf, last visited
2009.
[2] Bartlett M., Movellan J., and Sejnowski T.,
“Face Recognition by Independent Component
Analysis,” IEEE Transactions on Neural
Networks, vol. 13, no. 6, pp. 1450-1464, 2002.
[3] Breitenstein M., Kuettel D., Weise T., Van-Gool
L., and Pfister H., “Real-Time Face Pose
Estimation from Single Range Images,” in
Proceedings of IEEE Conference on Computer
Vision and Pattern Recognition, Anchorage,
Alaska, pp. 1-8, 2008.
[4] Dai M. and Zhou M., “On Automatic Human
Face Recognition,” Advances Biometrics, vol. 1,
pp. 41-48, 2003.
[5] Dai D. and Yuen P., “Wavelet Based
Discriminant Analysis for Face Recognition,”
Applied Mathematics and Computation, vol.
175, no. 1, pp. 307-318, 2006.
[6] Er M., Chen W., and Wu S., “High-Speed Face
Recognition Based on Discrete Cosine
Transform and RBF Neural Networks,” IEEE
Transactions on Neural Networks, vol. 16, no. 3,
pp. 679-691, 2005.
[7] Fan Z. and Lu B., “Fast Recognition of MultiView Faces with Feature Selection,” in
Proceedings of the 10th IEEE International
Conference on Computer Vision, vol. 1, pp. 7681, 2005.
[8] Fan Z. and Lu B., “Multi-View Face
Recognition with Min-Max Modular Support
Vector Machines,” in Proceedings of the 1st
International Conference on Advances in
Natural Computation, Changsha, China, vol.
3611, pp. 396-399, 2007.
[9] Gan J. and Zhang Y., “A New Approach for
Face Recognition Based on Singular Value
Features and Neural Networks,” Acta
Electronica Sinica, vol. 32, no. 1, pp. 56-58,
2004.
[10] Gan J., Zhang Y., and Mao S., “Adaptive
Principal Components Extraction Algorithm and
its Applications in the Feature Extraction of
Human Face,” Acta Electronica Sinica, vol. 30,
no. 7, pp. 1013-1016, 2002.
394
The International Arab Journal of Information Technology, Vol. 11, No. 4, July 2014
[11] Gao X. and Tian C., “Multi-View Face
Recognition Based on Tensor Subspace Analysis
and View Manifold Modeling,” Neurocomputing
Journal, vol. 72, no. 16-18, pp. 37423750, 2009.
[12] Gu S., Tan Y., and He X., “Laplacian Smoothing
Transform for Face Recognition,” Science in
China Series F-Information Sciences, vol. 53, no.
12, pp. 2415-2428, 2009.
[13] Gu S., Tan Y., and He X., “Discriminant Analysis
via Support Vectors,” Neurocomputing, vol. 73,
no. 10–12, pp. 1669-1675, 2009.
[14] Kanan H. and Moradi M., “A Genetic Algorithm
based Method for Face Localization and pose
Estimation,” in Proceedings of Sciences of
Electronic Technologies of Information and
Telecommunications, Tunisia, pp. 27-31, 2005.
[15] Kisku D., Mehrotra H., Sing J., and Gupta P.,
“SVM-based Multiview Face Recognition by
Generalization of Discriminant Analysis,”
International Journal of Intelligent Systems and
Technologies 3, vol. 2, no. 8, pp. 703-708, 2009.
[16] Kriegman D., “Eigenfaces vs. Fisherfaces:
Recognition Using Class Specic Linear
Projection,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 19, no. 7,
PP. 711-720, 1996.
[17] Kyperountas M., Tefas A., and Pitas I.,
“Weighted Piecewise LDA for Solving the Small
Sample Size Problem in Face Verication,” IEEE
Transactions on Neural Networks, vol. 18, no. 2,
pp. 506-519, 2007.
[18] Lee C. and Elgammal A., “Modeling View and
Posture Manifolds for Tracking,” in Proceedings
of IEEE 11th International Conference on
Computer Vision, Rio de Janeiro, Brazil, pp. 1-8,
2007.
[19] Liu X., Yang F., Paindavoine M., and Dang J.,
“Multi-Views Face Recognition System,” in
Proceedings of Colloque GRETSI, pp. 865-868,
2007.
[20] Lu J., Plataniotis K., and Venetsanopoulos A.,
“Face Recognition Using Kernel Direct
Discriminant Analysis Algorithms,” IEEE
Transactions on Neural Networks, vol. 14, no. 1,
pp. 117-126, 2003.
[21] Lu J., Plataniotis K., and Venetsanopoulos A., “A
Kernel Machine Based Approach for Multi-View
Face
Recognition,”
in
Proceedings
of
International Conference on Image Processing,
vol. 1, pp. 265-268, 2002.
[22] Lu B., Wang K., Utiyama M., and Isahara H., “A
Part-Versus-Part Method for Massively Parallel
Training of Support Vector Machines,” in
Proceedings of IEEE International Joint
Conference on Neural Networks, vol. 1, 2004.
[23] Morency L., Sundberg P., and Darrell T., “Pose
Estimation Using 3D View-Based Eigenspaces,”
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
in Proceedings of Analysis and Modeling of
Faces and Gestures, pp.45-52, 2003.
Murtaza M., Sharif M., Raza M., and Shah J.,
“Face Recognition Using Adaptive Margin
Fisher’s Criterion and Linear Discriminant
Analysis (AMFC-LDA),” the International Arab
Journal of Information Technology, vol. 11, no.
2, pp. 149-158, 2012.
Pentland A., Moghaddam B., and Starner T.,
“View-Based and Modular Eigenspaces for Face
Recognition,” in Proceedings of IEEE Computer
Society Conference on Computer Vision and
Pattern Recognition, Seattle, USA, pp. 84-91,
1994.
Rifkin R. and Klautau A., “In Defense of OneVs-All Classification,” Journal of Machine
Learning Research, vol. 5, pp. 101-141, 2004.
Scholkopf B., Smola A., Li X., and Muller K.,
“Nonlinear Component Analysis as a Kernel
Eigenvalue Problem,” Neural Computation, vol.
10, no. 5, pp. 1299-1319, 1998.
Tian C., Fan G., and Gao X., “Multi-View Face
Recognition
by
Non-Linear
Tensor
Decomposition,” in Proceedings of the 19th
International
Conference
on
Pattern
Recognition, Tampa, USA, pp. 1-4, 2008.
Turk M. and Pentland A., “Eigenfaces for
Recognition,”
Journal
of
Cognitive
Neuroscience, vol. 3, no. 1, pp. 71-86, 1991.
Vapnik V., The Nature of Statistical Learning
Theory, Springer Verlag, New York, USA, 1995.
Vasilescu M. and Terzopoulos D., “Multilinear
Image Analysis for Facial Recognition,” in
Proceedings of the 16th International Conference
on Pattern Recognition, vol. 2, pp. 511-514,
2002.
Yu M., Yan G., and Zhu Q., “New Face
Recognition Method Based on DWT/DCT
Combined Feature Selection,” in Proceedings of
International Conference Machine Learning and
Cybernetics, Dalian, China, pp. 3233-3236,
2006.
Zhang T., Tao D., Li X., and Yang J., “Patch
Alignment for Dimensionality Reduction,” IEEE
Transactions on Knowledge and Data
Engineering, vol. 21, no. 9, pp. 1299-1313,
2009.
Zhou D. and Yang X., “Face Recognition Using
Enhanced Sher Linear Discriminant Model with
Facial Combined Feature,” in Proceedings of the
8th Pacific Rim International Conference on
Artificial Intelligence, Auckland, New Zealand,
pp. 769-777, 2004.
Inter-Communication Classification for Multi-View Face Recognition
Chouaib Moujahdi is a PhD student
in Engineering Sciences Department
at Mohammed V-Agdal University.
He received his Master’s in 2010
from the same University. He had a
Fulbright Joint Supervision award to
visit the University of Nevada - Reno
for the period between 2012 and 2013. His research
interests include biometrics and pattern recognition. His
current work focuses on biometric security protection.
Sanaa Ghouzali received her
Master's and the Ph.D. degrees in
computer
science
and
telecommunications from Mohamed
V-Agdal University in 2004 and
2009, respectively. She
was a
Fulbright visiting student at Cornell
University between 2005 and 2007. She was an
assistant professor at ENSA (the National school of
Applied Sciences), within the University Abdelmalek
Essaadi, between 2009 and 2011. In 2012, she joined
the College of Computer and Information Sciences at
King Saud University where she is an assistant
professor in the Department of Information
Technology. Her research interests include statistical
pattern detection and recognition, biometrics, biometric
security and protection.
Mounia Mikram is a an assistant professor of
computer sciences and mathematics at the School of
Information Sciences, Rabat since 2010. She received
her master degree from Mohammed V University Rabat
(2003) and her PhD degree from Mohammed V
University, Rabat, and Bordeaux I University (2008).
Her research interests include pattern recognition,
computer vision, and biometrics security systems.
395
Abdul Wadood received his BE
degree from COMSATS Institute of
Information Technology, Islamabad,
Pakistan, in 2004. He did Masters
from University of Limoges, France
in 2007, and PhD in signal and
image processing from University
of Poitiers, Poitiers, France in 2011. Currently, he is
working as an assistant professor at the Department of
Computer Engineering, CCIS, King Saud University,
Riyadh, Saudi Arabia. His research interests are
focused on color image watermarking, steganography,
fingerprinting, and biometric template protection.
Mohammed Rziza Received his
national Doctorate in engineering
sciences,
image
processing
specialty, from the Faculty of
Science of the Mohammed V-Agdal
University, Rabat, Morocco, in
2002. He joined the Faculty of
Science, Rabat, Morocco, in 2003, as an assistant
professor. Since 1997, he is a member of the GSCM
group. His research interests include image
processing, pattern recognition, and stereovision.

Documents pareils