SlideShare ist ein Scribd-Unternehmen logo
1 von 6
Downloaden Sie, um offline zu lesen
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520

RESEARCH ARTICLE

www.ijera.com

OPEN ACCESS

Performance Parameter Analysis of Face Recognition Based On
Fuzzy C-Means Clustering, Shape and Corner Detection
Minj Salen Kujur1, Prof. Prashant Jain2
Department of Electronics & Communication Engineering college Jabalpur

Abstract
The aim of this research is to solve face recognition problem in area of biometric by using CSC hybrid method.
In our work first we make the cluster of face key points and parallely apply shape and corner method for
detection boundary of face and the corner of face objects. Our work is performing both face identification and
recognition on large face database. It is better than other previous work because our FAR, FRR and EER less
than. Also we calculate the recognition time in our work.
Keywords- Face Recognition, Clustering, Shape Descriptor, Corner Detection, RGB Image, Image Processing,
Color Model, Binary Image, 4-Connected Component.

I.

Introduction

Image of outdoor scenes are Now a day’s
face recognition is play very crucial role in recent
technology like mostly companies are adopting the
biometric identification for login but when images are
degraded than the performance of our system is
reduced. This study gives some idea about steps by
steps face recognition algorithm how face is
recognized but when quality of an image is degrade
due to some noise or any external reason than
matching process will not give accurate result for this
reason we adopt some restoration and enhancement
techniques like retinex theory for degrade image to
improve quality for better performance in next part of
my work.
Pattern recognition is the scientific
discipline whose goal is the classification of objects
into a number of categories or classes [1]. Depending
on the application, these objects can be images or
signal waveforms or any type of measurements that
need to be classified. We will refer to these objects
using the generic term patterns. Pattern recognition
has a long history, but before the 1960s it was mostly
the output of theoretical research in the area of
statistics [6].
We have studied a face recognition system
using the Principal Component Analysis (PCA)
algorithm with Euclidean distance as a classifier and
secondly Linear Discriminates Analysis (LDA) with
Euclidean distance as a classifier. Face recognition
systems try to find the identity of a given face image
according to their memory. The memory of a face
recognizer is generally simulated by a training set.
Independent Component Analysis (ICA) is similar to
PCA except that the distribution of the components is
designed to be non-Gaussian. Maximizing nonGaussianity promotes statistical independence.
The problem of face recognition can be stated as
follows: Given still images or video of a scene,
www.ijera.com

identifying one or more persons in the scene by using
a stored database of faces [3]. The problem is mainly
a classification problem. Training the face recognition
system with images from the known individuals and
classifying the newly coming test images into one of
the classes is the main aspect of the face recognition
systems.

II.

Face Recognition

The face plays a major role in our social
intercourse in conveying identity and emotion. Face
Recognition (FR) [1,3] is a challenging task and has
been one of the most successful applications of image
analysis and understanding in many fields such as
computer vision, pattern recognition. Image-based
face recognition techniques can be divided into two
groups according to the face representation which
they use, which being the appearance-based approach
and the feature-based approach, among which the
appearance-based is more popular, that use holistic
texture features [2,7].
With automatic face
recognition there are many applications in human
computer interaction, biometrics and security, etc.
Over the decades, many computer systems that can
recognize faces have been developed, some of which
have been in commercial use. Generally, appearancebased face recognition techniques are finished with
image matching in the space of compressed image. If
image matching done in the original space, it will
result in the curse of dimensionality in addition to the
problems of large computational complexity and
memory Face recognition has received considerable
interest as a widely accepted biometric, because of the
ease of collecting samples of a person, with or
without subject’s intension . Face recognition [10,11]
refers to an automated or semi automated process of
matching facial images. This type of technology
constitutes a wide group of technologies which all
work with face but use different Scanning techniques.
515 | P a g e
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520
Most common by far is 2D face recognition which is
easier and less expensive compared to the other
approaches.
There are four steps in face recognition process as
shown in flow chart:
1. Acquiring a sample: In a complete, full
implemented biometric system, a sensor takes an
observation. The sensor might be a camera and the
observation is a snapshot picture. In our system, a
sensor will be ignored, and a 2D face picture [8, 11]
“observation” will supplied manually.

III.

4. Declaring a Match: The face recognition system
will return a candidate match list of potential matches.
In this case, the intervention of a human operator will
be required in order to select the best fit from the
candidate list. An illustrative analogy is that of a
walk-through metal detector, where if a person causes
the detector to beep, a human operator steps in and
checks the person manually or with a hand-held
detector.
Step:1

Manual supply 2D face picture

Step:2

previous Work

A. Facial-Image Acquisition
In our research, original images were
obtained using a charge coupled devices (CCD)
camera with image dimensions of 384
243 pixels
encoded using 256 gray-scale levels.
In image acquisition, the subject sits 2.5 m
away from a CCD camera. On each site of the
camera, two 200-W lamps are placed at 30 angles to
the camera horizontally. The original images are
shown in Fig. 2.

2. Extracting Features: For this step, the relevant
data is extracted from the predefined captured sample.
This is can be done by the use of software where
many algorithms are available. The outcome of this
step is a biometric template which is a reduced set of
data that represents the unique features of the enrolled
user's face.
3. Comparison Templates: This depends on the
application at hand. For identification purposes, this
step will be a comparison between a given picture for
the subject and all the biometric templates stored on a
database. For verification, the biometric template of
the claimed identity will be retrieved (either from a
database or a storage medium presented by the
subject) and this will be compared to a given picture.

www.ijera.com

Fig. 2. Original face image.
B. Lighting Compensation
We adjusted the locations of the lamps to
change the lighting conditions. The total energy of an
image is the sum of the squares of the intensity
values. The average energy of all the face images in
the database is calculated. Then, each face image is
normalized to have energy equal to the average
energy
Energy

Intensity

(1)

C. Facial-Region Extraction
We adopt the face-detection method
presented in the method of detecting and extracting
the facial features in a gray-scale image is divided
into two stages. First, the possible human eye regions
are detected by testing all the valley regions in an
image. A pair of eye candidates is selected by means
of the genetic algorithm to form a possible face
candidate. In our method, a square block is used to
represent the detected face region. 3 shows an
example of a selected face region based on the
location of an eye pair. The relationships between the
eye pair and the face size are defined as follows:

Extracting Features
Step:3

Comparison of templates

Step:4

Declaring a Match
Fig. 1. Flow chart of the face detection

www.ijera.com

516 | P a g e
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520

www.ijera.com

Fig. 5. Extracted face image.
Fig. 3. Geometry of our head model.
Then, the symmetrical measure of the face is
calculated. The nose centerline (the perpendicular
bisector of the line linking the two eyes) in each facial
image is calculated. The difference between the left
half and right half from the nose centerline of a face
region should be small due to its symmetry. the
symmetrical measure is less than a threshold value,
the face candidate will be selected for further
verification.
After measuring the symmetry of a face
candidate, the existences of the different facial
features are also verified. The positions of the facial
features are verified by analyzing the projection of the
face candidate region. The facial feature regions will
exhibit a low value on the projection. A face region is
divided into three parts, each of which contains the
respective facial features. The -projection is the
average of gray-level intensities along each row of
pixels in a window. In order to reduce the effect of
the background in a face region, only the white
windows, as shown in Fig. 4, are considered in
computing the projections. The top window should
contain the eyebrows and the eyes, the middle
window should contain the nose, and the bottom
window should contain the mouth. When a face
candidate satisfies the aforementioned constraints, it
will be extracted as a face region. The extracted face
image is shown in Fig. 5.

D. Principal Component Analysis (PCA)
Let a pattern
be a two-dimensional (2-D)
array of intensity values. A pattern may also
be considered as a vector of dimension
. Denote
the database of patterns by
. Define the covariance matrix as follows [4].

Proposed Work
We present the model of the our work is given below
Input
Image

Enhance &
RGB to Y,Cb,Cr

Apply SD &
CD Method
Apply filtering.

Calculate
Corner

.

Other key
.
information

DATA
BASE
Merg
-ing

Com
pare

Show
Identification

Recogniti
on

Fig. 4. Windows for facial feature extraction.

www.ijera.com

Calculate
Clustering

Fig.6. Block diagram of proposed method

517 | P a g e
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520
Algorithm
Step1. We take query image.

IV.

Step2. Image enhancement (Filtering)
Step3. Convert into Y, Cb, Cr model.
Step4. By this process (step 3), we get three picture of
this query image (luminance, chrome blue & chrome
red)
Step5. Then we apply shape discripter on luminance
image.
Step6. Now we apply colour discripter on Cb(chrome
blue) &Cr.
Step7. Merging of result no.5&6 steps.
Step7. By step 7 we detect the face.
Step8. Now we save the detected shape in variables
(For subsequent analysis) c o m p l e t e
Step9. Prepare predefind database of 40 image.
Step10. Now we apply step 2-9 on prepare database
image.
Step11. Now we apply C-Means clustering
9& step 10 (parallely)

on step

www.ijera.com

FUZZY CLUSTERING

The “clusters” are functions that assign to
each object a number between zero and one, which is
called the membership of the object in the cluster.
Objects which are similar to each other are identified
by having high membership degrees in the same
cluster. It is also assumed that the membership
degrees are chosen so that their sum for each object is
one therefore, uzzy clustering is also a partition of the
set of objects. The most widely used fuzzy clustering
algorithm is the FCM algorithm.
FCM
FCM is a data clustering algorithm in which
each data point is associated with a cluster through a
membership degree. This technique divides a
collection N of data points into 𝑟𝑐 fuzzy groups and
finds a cluster center in each group such that a cost
function of a dissimilarity measure is minimized. The
algorithm employs fuzzy partitioning such that a
given data point can belong to several groups with a
degree specified by member membership grades
between 0 and 1. A fuzzy 𝑟𝑐 –partition of input feature
vector𝑋 = {𝑥1, 𝑥2, … … … … … 𝑥𝑁} is represented by
a matrix𝑈 = [𝜇 𝑖𝑘 ], and X is an N -element set of tdimensional vectors, each
representing a 32dimensional vector. The entries satisfy the following
constraints:
1≤ i ≤ 𝑟𝑐

𝜇 𝑖𝑘 𝜖 0,1 ,
𝑟𝑐
𝑖=1

𝜇 𝑖𝑘 =1,

1≤ i ≤ N

1≤ i ≤ N

(1)
(2)

Step12. If no. of clusters, clusters position, cluster
size are same then show recognition done and also
show recognize person name.

0˂

Step13. Determine no. of detected & no. of non
face.Image will be count & then calculate FAR,FRR
& EER.

𝑥 𝑘 (𝑘 = 1, … … … . 𝑁)
represents
the
feature
coordinate of the ith data. 𝜇 𝑖𝑘 is the membership
degree o 𝑥 𝑘 f to cluster . A proper partition 𝑈 𝑜𝑓 𝑋
may be defined by the minimization of the following
cost function:

Shape
In the proposed methodology face feature
extraction is implemented by using rectangle shape.
Applications of Corner Detectors
The use of interest points (and thus corner
detectors) to find corresponding points across
multiple images is a key step in many image
processing and computer vision applications. Some of
the most notable examples are:
· stereo matching
· image registration (of particular importance in
medical imaging)
· stitching of panoramic photographs
· object detection/recognition
· motion tracking
· robot navigation

𝐽 𝑚 𝑈, 𝐶

1≤ i ≤ 𝑟𝑐

𝜇 𝑖𝑘 ≤ 𝑁,

𝑁
𝑘=1

=

𝑟𝑐
𝑖=1

𝜇 𝑖𝑘

𝑚

𝑑2
𝑖𝑘

(3)

(4)

where𝑚 ∈ (1, ∞) is a weighting exponent, called a
fuzzifier,that is chosen according to the case.
When 𝑚 → 1 , the process
converges to a generalized classical -means.
When 𝑚 → ∞ , all clusters tend towards the center of
gravity of the whole data set. That is, the partition
becomes fuzzier with increasing 𝑚 .
𝐶 = [𝑐1, 𝑐2, … … … … . 𝑐 𝑟 𝑐 ] is the vector of the cluster
centers, and 𝑑 𝑖𝑘 is the distance between 𝑥 𝑘 and the ith
cluster. Bezdek [18] proved that if 𝑚 ≥ 1, 𝑑 2 > 0
𝑖𝑘
and 1≤ i ≤
𝑟𝑐 ,than U and C minimize
𝐽 𝑚 𝑈, 𝐶 only if their entries are computed as
𝜇∗ =
𝑖𝑘

www.ijera.com

𝑁
𝑘=1

1
𝑟𝑐
𝑖=1

𝑑 𝑖𝑘

(5)

2/( 𝑚 −1)

𝑑 𝑗𝑘

518 | P a g e
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520
𝐶 𝑖∗ =

𝑁
𝑚 𝑥
𝑘
𝑘=1 𝜇 𝑖𝑘
𝑁
𝜇 𝑖𝑘 𝑚
𝑘=1

.

www.ijera.com

(6)

One of the major factors that influence the
determination of appropriate clusters of points is the
dissimilarity measure chosen for the problem. Indeed,
the computation of the membership degree 𝜇 ∗
𝑖𝑘
depends on the definition of the distance
measure 𝑑 𝑖𝑘 , which is the inner product norm
(quadratic norm).
The squared quadratic norm (distance) between a
pattern vector 𝑥 𝑘 and the center 𝑐 𝑖 of the ith cluster
is defined as
𝑑2 =
𝑖𝑘

𝑥 𝑘 − 𝑐𝑖

2
𝐺

=

𝑥 𝑘 − 𝑐𝑖

𝑇

𝐺 𝑥 𝑘 − 𝑐𝑖

(7)

Fig.7 Query image of dataset

where𝐺 is any positive–definite matrix. The identity
matrix is the simplest and most popular choice for 𝐺.

V.

Algorithm of the Facial Images by
FCM

The FCM algorithm consists of a series of
iterations using (5) and (6). This algorithm converges
to a local minimum point of 𝐽 𝑚 𝑈, 𝐶 . We use the
FCM as follows to determine the cluster centers 𝑐 𝑖
and the membership matrix 𝑈 .
Step 1)
Initially, the membership matrix is
constructed using random values between 0 and 1
such that constraints (1), (2), and (3) are satisfied.
Step 2) The membership function is computed as
follows.
a) For each cluster 𝑖 , the fuzzy cluster center 𝑐 𝑖 is
computed using (6).
b) All cluster centers which are too close to each
other are eliminated.
For each cluster 𝑖 , the distance𝐷 𝑖𝑘 is computed
As 𝐷 𝑖𝑘 = 𝑐 𝑖 − 𝑐 𝑘 .
When 𝐷 𝑖𝑘 ≤ 𝑎𝑣𝑒 𝑑 , 𝜇 𝑘𝑗 = 0for j=1 to N ,
where 𝑎𝑣𝑒 𝑑 is the average distance value.
c) For each cluster 𝑖 , the distance 𝑑 𝑖𝑘 is computed
using (7)
d) The cost function using (4) is computed. Stop
if its improvement over the previous iteration
below a threshold.
e) A new using (4) is computed and Step 2) is
repeated.

VI.

Fig.9. a)matching with person name.
b)matching done and recognition time

Experimental Result

Below figure show the GUI of our work, that
getting the face image for analysis. Next figure
calculate the only face of the face image. After that
we apply our methodology on this query image and
finally archive the matching face image, name of
image person, clustering on face image and diagram
of FAR,FRR.

www.ijera.com

Fig.8. Find face of image with only face
and luminance image

Fig.10. Graph of FAR and FRR

519 | P a g e
Minj Salen Kujur et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520
[7]

[8]

[9]

[10]
Fig.11. Graph of recognition rate

VII.

Conclusion

In this paper, we have developed a hybrid
method that is very effectively work on face image. It
selects a face image then find the only face image and
matching with respect to database. It reduces the
deficiency of existing methods like PCA, correlation
based etc. This hybrid method gives better result than
all the other individual method. In this work we
calculate FAR and FRR. In future we add some other
concept like 2D cross correlation and moment
invariants of face, with this approach and get a very
good result for fingerprint matching.

[11]

[12]

[13]

[14]

Refference
[1]

[2]

[3]

[4]

[5]

[6]

R. Brunelli, T. Poggio, Face recognition:
features versus templates,IEEE Transactions
on Pattern Analysis and Machine
Intelligence 15(10) (1993) 1042–1052.
M. Turk, A. Pentland, Eigenfaces for
recognition, Journal of Cognitive Science
1991; 71–86.
P.N. Belhumeur, J.P. Hespanha, D.J.
Kriegman, Eigenfaces vs. fisherfaces:
recognition using class specific linear
projection, IEEE Transactions on Pattern
Analysis and Machine Intelligence 19 (7)
(1997) 711–720.
M.S. Bartlett, H.M. Lades, T.J. Sejnowski,
Independent component representations for
face recognition, Proceedings of the SPIE
Conference on Human Vision and Electronic
Imaging III, San Jose, CA, 1998 pp. 528–
539.
Y. Moses, Y. Adini, S. Ullman, Face
recognition: the problem of compensating
for changes in illumination direction, IEEE
Transactions on Pattern Analysis and
Machine Intelligence 19 (7) (1997) 721–732.
G.C. Feng, P.C. Yuen, D.Q. Dai, Human
face recognition using PCA on wavelet
subband, SPIE Journal of Electronic Imaging
9 (2) (2000) 226–233.

www.ijera.com

[15]

[16]

[17]

[18]

[19]

www.ijera.com

J. Tang, R. Nakatsu, S. Kawato, J. Ohya, A
wavelet-transform based asker identification
system for smart multi-point teleconferences, Journal of the Visualization
Society of Japan 20 (1) (2000) 303–306.
C. Garcia, G. Zikos, G. Tziritas, Wavelet
packet analysis for face recognition, Image
and Vision Computing 18 (4) (2000) 289–
297.
B. Li, Y. Liu, When eigenfaces are
combined with wavelets, International
Journal of Knowledge-Based Systems 15
(5/6) (2002) 343–347.
J.T. Chien, C.C. Wu, Discriminant
waveletfaces and nearest feature classifiers
for face recognition, IEEE Transactions on
Pattern Analysis and Machine Intelligence
24 (12) (2002) 1644–1649.
P.J. Phillips, Matching pursuit filters applied
to face identification, IEEE Transactions on
Image Processing 7 (8) (1998) 1150–1164.
A. Hyva¨rinen, E. Oja, Independent
component
analysis:
algorithms
and
applications, Neural Networks 13 (2000)
411–430.
J. Kittler, M. Hatef, R.P. Duin, J.G. Matas,
On combining classifiers, IEEE Transactions
on Pattern Analysis and Machine
Intelligence 20 (3) (1998) 226–239.
I. Daubechies, Ten lectures on wavelets,
CBMS-NSF Regional Conference Series in
Applied Mathematics, vol. 61, SIAM Press,
Philadelphia, 1992.
T. Sim, S. Baker, M. Bsat, TheCMUpose,
illumination, and expression (PIE) database,
Proceedings of the IEEE International
Conference on Automatic Face and Gesture
Recognition, May 2002.
P.J. Phillips, H. Moon, S.A. Rizvi, P.J.
Rauss, The FERET evaluation methodology
for face-recognition algorithms, IEEE
Transactions on Pattern Analysis and
Machine Intelligence 22 (10) (2000) 1090–
1104.
K. W. Wong, K. M. Lam, and W. C. Siu,
“An efficient algorithm for face detection
and facial feature extraction under different
conditions,”Pattern Recognit., vol. 34, no.
10, pp. 1993–2004, 2001.
J. Bezdek, “A convergence theorem for the
fuzzy ISODATA clustering algorithms,”
IEEE Trans. Pattern Anal. Mach. Intell., vol.
PAMI-2, no.1, pp. 1–8, Jan. 1981.
W. Pedrycz, “Conditional fuzzy clustering in
the design of radial basis function neural
networks,” IEEE Trans. Neural Netw., vol.
9, no. 4, pp.601–612, Jul. 1998.

520 | P a g e

Weitere ähnliche Inhalte

Was ist angesagt?

Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
Editor IJARCET
 
Facial expression recongnition Techniques, Database and Classifiers
Facial expression recongnition Techniques, Database and Classifiers Facial expression recongnition Techniques, Database and Classifiers
Facial expression recongnition Techniques, Database and Classifiers
Rupinder Saini
 

Was ist angesagt? (19)

Ch 1
Ch 1Ch 1
Ch 1
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Ch 2
Ch 2Ch 2
Ch 2
 
IRJET - A Review on Face Recognition using Deep Learning Algorithm
IRJET -  	  A Review on Face Recognition using Deep Learning AlgorithmIRJET -  	  A Review on Face Recognition using Deep Learning Algorithm
IRJET - A Review on Face Recognition using Deep Learning Algorithm
 
FACE DETECTION USING PRINCIPAL COMPONENT ANALYSIS
FACE DETECTION USING PRINCIPAL COMPONENT ANALYSISFACE DETECTION USING PRINCIPAL COMPONENT ANALYSIS
FACE DETECTION USING PRINCIPAL COMPONENT ANALYSIS
 
Asymmetrical Half-Join Method on Dual Vision Face Recognition
Asymmetrical Half-Join Method on Dual Vision Face Recognition Asymmetrical Half-Join Method on Dual Vision Face Recognition
Asymmetrical Half-Join Method on Dual Vision Face Recognition
 
FUSION BASED MULTIMODAL AUTHENTICATION IN BIOMETRICS USING CONTEXT-SENSITIVE ...
FUSION BASED MULTIMODAL AUTHENTICATION IN BIOMETRICS USING CONTEXT-SENSITIVE ...FUSION BASED MULTIMODAL AUTHENTICATION IN BIOMETRICS USING CONTEXT-SENSITIVE ...
FUSION BASED MULTIMODAL AUTHENTICATION IN BIOMETRICS USING CONTEXT-SENSITIVE ...
 
Facial expression recognition based on image feature
Facial expression recognition based on image featureFacial expression recognition based on image feature
Facial expression recognition based on image feature
 
Face detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategyFace detection for video summary using enhancement based fusion strategy
Face detection for video summary using enhancement based fusion strategy
 
A Hybrid Approach to Recognize Facial Image using Feature Extraction Method
A Hybrid Approach to Recognize Facial Image using Feature Extraction MethodA Hybrid Approach to Recognize Facial Image using Feature Extraction Method
A Hybrid Approach to Recognize Facial Image using Feature Extraction Method
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and Development
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
 
Aa4102207210
Aa4102207210Aa4102207210
Aa4102207210
 
Multi modal face recognition using block based curvelet features
Multi modal face recognition using block based curvelet featuresMulti modal face recognition using block based curvelet features
Multi modal face recognition using block based curvelet features
 
A study on face recognition technique based on eigenface
A study on face recognition technique based on eigenfaceA study on face recognition technique based on eigenface
A study on face recognition technique based on eigenface
 
IRJET- Digiyathra
IRJET-  	  DigiyathraIRJET-  	  Digiyathra
IRJET- Digiyathra
 
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
 
IRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern AlgorithmIRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
IRJET- A Review on Face Recognition using Local Binary Pattern Algorithm
 
Facial expression recongnition Techniques, Database and Classifiers
Facial expression recongnition Techniques, Database and Classifiers Facial expression recongnition Techniques, Database and Classifiers
Facial expression recongnition Techniques, Database and Classifiers
 

Andere mochten auch

Andere mochten auch (8)

50 Essential Content Marketing Hacks (Content Marketing World)
50 Essential Content Marketing Hacks (Content Marketing World)50 Essential Content Marketing Hacks (Content Marketing World)
50 Essential Content Marketing Hacks (Content Marketing World)
 
Prototyping is an attitude
Prototyping is an attitudePrototyping is an attitude
Prototyping is an attitude
 
10 Insightful Quotes On Designing A Better Customer Experience
10 Insightful Quotes On Designing A Better Customer Experience10 Insightful Quotes On Designing A Better Customer Experience
10 Insightful Quotes On Designing A Better Customer Experience
 
Learn BEM: CSS Naming Convention
Learn BEM: CSS Naming ConventionLearn BEM: CSS Naming Convention
Learn BEM: CSS Naming Convention
 
How to Build a Dynamic Social Media Plan
How to Build a Dynamic Social Media PlanHow to Build a Dynamic Social Media Plan
How to Build a Dynamic Social Media Plan
 
SEO: Getting Personal
SEO: Getting PersonalSEO: Getting Personal
SEO: Getting Personal
 
Lightning Talk #9: How UX and Data Storytelling Can Shape Policy by Mika Aldaba
Lightning Talk #9: How UX and Data Storytelling Can Shape Policy by Mika AldabaLightning Talk #9: How UX and Data Storytelling Can Shape Policy by Mika Aldaba
Lightning Talk #9: How UX and Data Storytelling Can Shape Policy by Mika Aldaba
 
Succession “Losers”: What Happens to Executives Passed Over for the CEO Job?
Succession “Losers”: What Happens to Executives Passed Over for the CEO Job? Succession “Losers”: What Happens to Executives Passed Over for the CEO Job?
Succession “Losers”: What Happens to Executives Passed Over for the CEO Job?
 

Ähnlich wie Ck36515520

Image–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlabImage–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlab
Ijcem Journal
 
Innovative Analytic and Holistic Combined Face Recognition and Verification M...
Innovative Analytic and Holistic Combined Face Recognition and Verification M...Innovative Analytic and Holistic Combined Face Recognition and Verification M...
Innovative Analytic and Holistic Combined Face Recognition and Verification M...
ijbuiiir1
 
Face recogntion using PCA algorithm
Face recogntion using PCA algorithmFace recogntion using PCA algorithm
Face recogntion using PCA algorithm
Ashwini Awatare
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
Editor IJARCET
 
Paper id 24201475
Paper id 24201475Paper id 24201475
Paper id 24201475
IJRAT
 

Ähnlich wie Ck36515520 (20)

Real time facial expression analysis using pca
Real time facial expression analysis using pcaReal time facial expression analysis using pca
Real time facial expression analysis using pca
 
A study of techniques for facial detection and expression classification
A study of techniques for facial detection and expression classificationA study of techniques for facial detection and expression classification
A study of techniques for facial detection and expression classification
 
Criminal Detection System
Criminal Detection SystemCriminal Detection System
Criminal Detection System
 
Model Based Emotion Detection using Point Clouds
Model Based Emotion Detection using Point CloudsModel Based Emotion Detection using Point Clouds
Model Based Emotion Detection using Point Clouds
 
Image–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlabImage–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlab
 
184
184184
184
 
Face Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCAFace Recognition Using Gabor features And PCA
Face Recognition Using Gabor features And PCA
 
Realtime human face tracking and recognition system on uncontrolled environment
Realtime human face tracking and recognition system on  uncontrolled environmentRealtime human face tracking and recognition system on  uncontrolled environment
Realtime human face tracking and recognition system on uncontrolled environment
 
76 s201920
76 s20192076 s201920
76 s201920
 
Face recognition system
Face recognition systemFace recognition system
Face recognition system
 
Innovative Analytic and Holistic Combined Face Recognition and Verification M...
Innovative Analytic and Holistic Combined Face Recognition and Verification M...Innovative Analytic and Holistic Combined Face Recognition and Verification M...
Innovative Analytic and Holistic Combined Face Recognition and Verification M...
 
Face Detection Using Modified Viola Jones Algorithm
Face Detection Using Modified Viola Jones AlgorithmFace Detection Using Modified Viola Jones Algorithm
Face Detection Using Modified Viola Jones Algorithm
 
Face recogntion using PCA algorithm
Face recogntion using PCA algorithmFace recogntion using PCA algorithm
Face recogntion using PCA algorithm
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
 
Face recognition
Face recognitionFace recognition
Face recognition
 
Real time voting system using face recognition for different expressions and ...
Real time voting system using face recognition for different expressions and ...Real time voting system using face recognition for different expressions and ...
Real time voting system using face recognition for different expressions and ...
 
Paper id 24201475
Paper id 24201475Paper id 24201475
Paper id 24201475
 
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
A FACE RECOGNITION USING LINEAR-DIAGONAL BINARY GRAPH PATTERN FEATURE EXTRACT...
 
A Face Recognition Using Linear-Diagonal Binary Graph Pattern Feature Extract...
A Face Recognition Using Linear-Diagonal Binary Graph Pattern Feature Extract...A Face Recognition Using Linear-Diagonal Binary Graph Pattern Feature Extract...
A Face Recognition Using Linear-Diagonal Binary Graph Pattern Feature Extract...
 
C017431730
C017431730C017431730
C017431730
 

Kürzlich hochgeladen

+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Victor Rentea
 

Kürzlich hochgeladen (20)

MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024
 
Spring Boot vs Quarkus the ultimate battle - DevoxxUK
Spring Boot vs Quarkus the ultimate battle - DevoxxUKSpring Boot vs Quarkus the ultimate battle - DevoxxUK
Spring Boot vs Quarkus the ultimate battle - DevoxxUK
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 

Ck36515520

  • 1. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 RESEARCH ARTICLE www.ijera.com OPEN ACCESS Performance Parameter Analysis of Face Recognition Based On Fuzzy C-Means Clustering, Shape and Corner Detection Minj Salen Kujur1, Prof. Prashant Jain2 Department of Electronics & Communication Engineering college Jabalpur Abstract The aim of this research is to solve face recognition problem in area of biometric by using CSC hybrid method. In our work first we make the cluster of face key points and parallely apply shape and corner method for detection boundary of face and the corner of face objects. Our work is performing both face identification and recognition on large face database. It is better than other previous work because our FAR, FRR and EER less than. Also we calculate the recognition time in our work. Keywords- Face Recognition, Clustering, Shape Descriptor, Corner Detection, RGB Image, Image Processing, Color Model, Binary Image, 4-Connected Component. I. Introduction Image of outdoor scenes are Now a day’s face recognition is play very crucial role in recent technology like mostly companies are adopting the biometric identification for login but when images are degraded than the performance of our system is reduced. This study gives some idea about steps by steps face recognition algorithm how face is recognized but when quality of an image is degrade due to some noise or any external reason than matching process will not give accurate result for this reason we adopt some restoration and enhancement techniques like retinex theory for degrade image to improve quality for better performance in next part of my work. Pattern recognition is the scientific discipline whose goal is the classification of objects into a number of categories or classes [1]. Depending on the application, these objects can be images or signal waveforms or any type of measurements that need to be classified. We will refer to these objects using the generic term patterns. Pattern recognition has a long history, but before the 1960s it was mostly the output of theoretical research in the area of statistics [6]. We have studied a face recognition system using the Principal Component Analysis (PCA) algorithm with Euclidean distance as a classifier and secondly Linear Discriminates Analysis (LDA) with Euclidean distance as a classifier. Face recognition systems try to find the identity of a given face image according to their memory. The memory of a face recognizer is generally simulated by a training set. Independent Component Analysis (ICA) is similar to PCA except that the distribution of the components is designed to be non-Gaussian. Maximizing nonGaussianity promotes statistical independence. The problem of face recognition can be stated as follows: Given still images or video of a scene, www.ijera.com identifying one or more persons in the scene by using a stored database of faces [3]. The problem is mainly a classification problem. Training the face recognition system with images from the known individuals and classifying the newly coming test images into one of the classes is the main aspect of the face recognition systems. II. Face Recognition The face plays a major role in our social intercourse in conveying identity and emotion. Face Recognition (FR) [1,3] is a challenging task and has been one of the most successful applications of image analysis and understanding in many fields such as computer vision, pattern recognition. Image-based face recognition techniques can be divided into two groups according to the face representation which they use, which being the appearance-based approach and the feature-based approach, among which the appearance-based is more popular, that use holistic texture features [2,7]. With automatic face recognition there are many applications in human computer interaction, biometrics and security, etc. Over the decades, many computer systems that can recognize faces have been developed, some of which have been in commercial use. Generally, appearancebased face recognition techniques are finished with image matching in the space of compressed image. If image matching done in the original space, it will result in the curse of dimensionality in addition to the problems of large computational complexity and memory Face recognition has received considerable interest as a widely accepted biometric, because of the ease of collecting samples of a person, with or without subject’s intension . Face recognition [10,11] refers to an automated or semi automated process of matching facial images. This type of technology constitutes a wide group of technologies which all work with face but use different Scanning techniques. 515 | P a g e
  • 2. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 Most common by far is 2D face recognition which is easier and less expensive compared to the other approaches. There are four steps in face recognition process as shown in flow chart: 1. Acquiring a sample: In a complete, full implemented biometric system, a sensor takes an observation. The sensor might be a camera and the observation is a snapshot picture. In our system, a sensor will be ignored, and a 2D face picture [8, 11] “observation” will supplied manually. III. 4. Declaring a Match: The face recognition system will return a candidate match list of potential matches. In this case, the intervention of a human operator will be required in order to select the best fit from the candidate list. An illustrative analogy is that of a walk-through metal detector, where if a person causes the detector to beep, a human operator steps in and checks the person manually or with a hand-held detector. Step:1 Manual supply 2D face picture Step:2 previous Work A. Facial-Image Acquisition In our research, original images were obtained using a charge coupled devices (CCD) camera with image dimensions of 384 243 pixels encoded using 256 gray-scale levels. In image acquisition, the subject sits 2.5 m away from a CCD camera. On each site of the camera, two 200-W lamps are placed at 30 angles to the camera horizontally. The original images are shown in Fig. 2. 2. Extracting Features: For this step, the relevant data is extracted from the predefined captured sample. This is can be done by the use of software where many algorithms are available. The outcome of this step is a biometric template which is a reduced set of data that represents the unique features of the enrolled user's face. 3. Comparison Templates: This depends on the application at hand. For identification purposes, this step will be a comparison between a given picture for the subject and all the biometric templates stored on a database. For verification, the biometric template of the claimed identity will be retrieved (either from a database or a storage medium presented by the subject) and this will be compared to a given picture. www.ijera.com Fig. 2. Original face image. B. Lighting Compensation We adjusted the locations of the lamps to change the lighting conditions. The total energy of an image is the sum of the squares of the intensity values. The average energy of all the face images in the database is calculated. Then, each face image is normalized to have energy equal to the average energy Energy Intensity (1) C. Facial-Region Extraction We adopt the face-detection method presented in the method of detecting and extracting the facial features in a gray-scale image is divided into two stages. First, the possible human eye regions are detected by testing all the valley regions in an image. A pair of eye candidates is selected by means of the genetic algorithm to form a possible face candidate. In our method, a square block is used to represent the detected face region. 3 shows an example of a selected face region based on the location of an eye pair. The relationships between the eye pair and the face size are defined as follows: Extracting Features Step:3 Comparison of templates Step:4 Declaring a Match Fig. 1. Flow chart of the face detection www.ijera.com 516 | P a g e
  • 3. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 www.ijera.com Fig. 5. Extracted face image. Fig. 3. Geometry of our head model. Then, the symmetrical measure of the face is calculated. The nose centerline (the perpendicular bisector of the line linking the two eyes) in each facial image is calculated. The difference between the left half and right half from the nose centerline of a face region should be small due to its symmetry. the symmetrical measure is less than a threshold value, the face candidate will be selected for further verification. After measuring the symmetry of a face candidate, the existences of the different facial features are also verified. The positions of the facial features are verified by analyzing the projection of the face candidate region. The facial feature regions will exhibit a low value on the projection. A face region is divided into three parts, each of which contains the respective facial features. The -projection is the average of gray-level intensities along each row of pixels in a window. In order to reduce the effect of the background in a face region, only the white windows, as shown in Fig. 4, are considered in computing the projections. The top window should contain the eyebrows and the eyes, the middle window should contain the nose, and the bottom window should contain the mouth. When a face candidate satisfies the aforementioned constraints, it will be extracted as a face region. The extracted face image is shown in Fig. 5. D. Principal Component Analysis (PCA) Let a pattern be a two-dimensional (2-D) array of intensity values. A pattern may also be considered as a vector of dimension . Denote the database of patterns by . Define the covariance matrix as follows [4]. Proposed Work We present the model of the our work is given below Input Image Enhance & RGB to Y,Cb,Cr Apply SD & CD Method Apply filtering. Calculate Corner . Other key . information DATA BASE Merg -ing Com pare Show Identification Recogniti on Fig. 4. Windows for facial feature extraction. www.ijera.com Calculate Clustering Fig.6. Block diagram of proposed method 517 | P a g e
  • 4. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 Algorithm Step1. We take query image. IV. Step2. Image enhancement (Filtering) Step3. Convert into Y, Cb, Cr model. Step4. By this process (step 3), we get three picture of this query image (luminance, chrome blue & chrome red) Step5. Then we apply shape discripter on luminance image. Step6. Now we apply colour discripter on Cb(chrome blue) &Cr. Step7. Merging of result no.5&6 steps. Step7. By step 7 we detect the face. Step8. Now we save the detected shape in variables (For subsequent analysis) c o m p l e t e Step9. Prepare predefind database of 40 image. Step10. Now we apply step 2-9 on prepare database image. Step11. Now we apply C-Means clustering 9& step 10 (parallely) on step www.ijera.com FUZZY CLUSTERING The “clusters” are functions that assign to each object a number between zero and one, which is called the membership of the object in the cluster. Objects which are similar to each other are identified by having high membership degrees in the same cluster. It is also assumed that the membership degrees are chosen so that their sum for each object is one therefore, uzzy clustering is also a partition of the set of objects. The most widely used fuzzy clustering algorithm is the FCM algorithm. FCM FCM is a data clustering algorithm in which each data point is associated with a cluster through a membership degree. This technique divides a collection N of data points into 𝑟𝑐 fuzzy groups and finds a cluster center in each group such that a cost function of a dissimilarity measure is minimized. The algorithm employs fuzzy partitioning such that a given data point can belong to several groups with a degree specified by member membership grades between 0 and 1. A fuzzy 𝑟𝑐 –partition of input feature vector𝑋 = {𝑥1, 𝑥2, … … … … … 𝑥𝑁} is represented by a matrix𝑈 = [𝜇 𝑖𝑘 ], and X is an N -element set of tdimensional vectors, each representing a 32dimensional vector. The entries satisfy the following constraints: 1≤ i ≤ 𝑟𝑐 𝜇 𝑖𝑘 𝜖 0,1 , 𝑟𝑐 𝑖=1 𝜇 𝑖𝑘 =1, 1≤ i ≤ N 1≤ i ≤ N (1) (2) Step12. If no. of clusters, clusters position, cluster size are same then show recognition done and also show recognize person name. 0˂ Step13. Determine no. of detected & no. of non face.Image will be count & then calculate FAR,FRR & EER. 𝑥 𝑘 (𝑘 = 1, … … … . 𝑁) represents the feature coordinate of the ith data. 𝜇 𝑖𝑘 is the membership degree o 𝑥 𝑘 f to cluster . A proper partition 𝑈 𝑜𝑓 𝑋 may be defined by the minimization of the following cost function: Shape In the proposed methodology face feature extraction is implemented by using rectangle shape. Applications of Corner Detectors The use of interest points (and thus corner detectors) to find corresponding points across multiple images is a key step in many image processing and computer vision applications. Some of the most notable examples are: · stereo matching · image registration (of particular importance in medical imaging) · stitching of panoramic photographs · object detection/recognition · motion tracking · robot navigation 𝐽 𝑚 𝑈, 𝐶 1≤ i ≤ 𝑟𝑐 𝜇 𝑖𝑘 ≤ 𝑁, 𝑁 𝑘=1 = 𝑟𝑐 𝑖=1 𝜇 𝑖𝑘 𝑚 𝑑2 𝑖𝑘 (3) (4) where𝑚 ∈ (1, ∞) is a weighting exponent, called a fuzzifier,that is chosen according to the case. When 𝑚 → 1 , the process converges to a generalized classical -means. When 𝑚 → ∞ , all clusters tend towards the center of gravity of the whole data set. That is, the partition becomes fuzzier with increasing 𝑚 . 𝐶 = [𝑐1, 𝑐2, … … … … . 𝑐 𝑟 𝑐 ] is the vector of the cluster centers, and 𝑑 𝑖𝑘 is the distance between 𝑥 𝑘 and the ith cluster. Bezdek [18] proved that if 𝑚 ≥ 1, 𝑑 2 > 0 𝑖𝑘 and 1≤ i ≤ 𝑟𝑐 ,than U and C minimize 𝐽 𝑚 𝑈, 𝐶 only if their entries are computed as 𝜇∗ = 𝑖𝑘 www.ijera.com 𝑁 𝑘=1 1 𝑟𝑐 𝑖=1 𝑑 𝑖𝑘 (5) 2/( 𝑚 −1) 𝑑 𝑗𝑘 518 | P a g e
  • 5. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 𝐶 𝑖∗ = 𝑁 𝑚 𝑥 𝑘 𝑘=1 𝜇 𝑖𝑘 𝑁 𝜇 𝑖𝑘 𝑚 𝑘=1 . www.ijera.com (6) One of the major factors that influence the determination of appropriate clusters of points is the dissimilarity measure chosen for the problem. Indeed, the computation of the membership degree 𝜇 ∗ 𝑖𝑘 depends on the definition of the distance measure 𝑑 𝑖𝑘 , which is the inner product norm (quadratic norm). The squared quadratic norm (distance) between a pattern vector 𝑥 𝑘 and the center 𝑐 𝑖 of the ith cluster is defined as 𝑑2 = 𝑖𝑘 𝑥 𝑘 − 𝑐𝑖 2 𝐺 = 𝑥 𝑘 − 𝑐𝑖 𝑇 𝐺 𝑥 𝑘 − 𝑐𝑖 (7) Fig.7 Query image of dataset where𝐺 is any positive–definite matrix. The identity matrix is the simplest and most popular choice for 𝐺. V. Algorithm of the Facial Images by FCM The FCM algorithm consists of a series of iterations using (5) and (6). This algorithm converges to a local minimum point of 𝐽 𝑚 𝑈, 𝐶 . We use the FCM as follows to determine the cluster centers 𝑐 𝑖 and the membership matrix 𝑈 . Step 1) Initially, the membership matrix is constructed using random values between 0 and 1 such that constraints (1), (2), and (3) are satisfied. Step 2) The membership function is computed as follows. a) For each cluster 𝑖 , the fuzzy cluster center 𝑐 𝑖 is computed using (6). b) All cluster centers which are too close to each other are eliminated. For each cluster 𝑖 , the distance𝐷 𝑖𝑘 is computed As 𝐷 𝑖𝑘 = 𝑐 𝑖 − 𝑐 𝑘 . When 𝐷 𝑖𝑘 ≤ 𝑎𝑣𝑒 𝑑 , 𝜇 𝑘𝑗 = 0for j=1 to N , where 𝑎𝑣𝑒 𝑑 is the average distance value. c) For each cluster 𝑖 , the distance 𝑑 𝑖𝑘 is computed using (7) d) The cost function using (4) is computed. Stop if its improvement over the previous iteration below a threshold. e) A new using (4) is computed and Step 2) is repeated. VI. Fig.9. a)matching with person name. b)matching done and recognition time Experimental Result Below figure show the GUI of our work, that getting the face image for analysis. Next figure calculate the only face of the face image. After that we apply our methodology on this query image and finally archive the matching face image, name of image person, clustering on face image and diagram of FAR,FRR. www.ijera.com Fig.8. Find face of image with only face and luminance image Fig.10. Graph of FAR and FRR 519 | P a g e
  • 6. Minj Salen Kujur et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 6, Nov-Dec 2013, pp.515-520 [7] [8] [9] [10] Fig.11. Graph of recognition rate VII. Conclusion In this paper, we have developed a hybrid method that is very effectively work on face image. It selects a face image then find the only face image and matching with respect to database. It reduces the deficiency of existing methods like PCA, correlation based etc. This hybrid method gives better result than all the other individual method. In this work we calculate FAR and FRR. In future we add some other concept like 2D cross correlation and moment invariants of face, with this approach and get a very good result for fingerprint matching. [11] [12] [13] [14] Refference [1] [2] [3] [4] [5] [6] R. Brunelli, T. Poggio, Face recognition: features versus templates,IEEE Transactions on Pattern Analysis and Machine Intelligence 15(10) (1993) 1042–1052. M. Turk, A. Pentland, Eigenfaces for recognition, Journal of Cognitive Science 1991; 71–86. P.N. Belhumeur, J.P. Hespanha, D.J. Kriegman, Eigenfaces vs. fisherfaces: recognition using class specific linear projection, IEEE Transactions on Pattern Analysis and Machine Intelligence 19 (7) (1997) 711–720. M.S. Bartlett, H.M. Lades, T.J. Sejnowski, Independent component representations for face recognition, Proceedings of the SPIE Conference on Human Vision and Electronic Imaging III, San Jose, CA, 1998 pp. 528– 539. Y. Moses, Y. Adini, S. Ullman, Face recognition: the problem of compensating for changes in illumination direction, IEEE Transactions on Pattern Analysis and Machine Intelligence 19 (7) (1997) 721–732. G.C. Feng, P.C. Yuen, D.Q. Dai, Human face recognition using PCA on wavelet subband, SPIE Journal of Electronic Imaging 9 (2) (2000) 226–233. www.ijera.com [15] [16] [17] [18] [19] www.ijera.com J. Tang, R. Nakatsu, S. Kawato, J. Ohya, A wavelet-transform based asker identification system for smart multi-point teleconferences, Journal of the Visualization Society of Japan 20 (1) (2000) 303–306. C. Garcia, G. Zikos, G. Tziritas, Wavelet packet analysis for face recognition, Image and Vision Computing 18 (4) (2000) 289– 297. B. Li, Y. Liu, When eigenfaces are combined with wavelets, International Journal of Knowledge-Based Systems 15 (5/6) (2002) 343–347. J.T. Chien, C.C. Wu, Discriminant waveletfaces and nearest feature classifiers for face recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence 24 (12) (2002) 1644–1649. P.J. Phillips, Matching pursuit filters applied to face identification, IEEE Transactions on Image Processing 7 (8) (1998) 1150–1164. A. Hyva¨rinen, E. Oja, Independent component analysis: algorithms and applications, Neural Networks 13 (2000) 411–430. J. Kittler, M. Hatef, R.P. Duin, J.G. Matas, On combining classifiers, IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (3) (1998) 226–239. I. Daubechies, Ten lectures on wavelets, CBMS-NSF Regional Conference Series in Applied Mathematics, vol. 61, SIAM Press, Philadelphia, 1992. T. Sim, S. Baker, M. Bsat, TheCMUpose, illumination, and expression (PIE) database, Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, May 2002. P.J. Phillips, H. Moon, S.A. Rizvi, P.J. Rauss, The FERET evaluation methodology for face-recognition algorithms, IEEE Transactions on Pattern Analysis and Machine Intelligence 22 (10) (2000) 1090– 1104. K. W. Wong, K. M. Lam, and W. C. Siu, “An efficient algorithm for face detection and facial feature extraction under different conditions,”Pattern Recognit., vol. 34, no. 10, pp. 1993–2004, 2001. J. Bezdek, “A convergence theorem for the fuzzy ISODATA clustering algorithms,” IEEE Trans. Pattern Anal. Mach. Intell., vol. PAMI-2, no.1, pp. 1–8, Jan. 1981. W. Pedrycz, “Conditional fuzzy clustering in the design of radial basis function neural networks,” IEEE Trans. Neural Netw., vol. 9, no. 4, pp.601–612, Jul. 1998. 520 | P a g e