Facial Recognition Using Scale Invariant Feature Transform
Through this project I'm trying to analyze the performance of different variations of SIFT. Then I will implement a solution using the most suited method for my scenario.
Tuesday, July 21, 2015
A Literature Review
Introduction
Humans are very good at recognizing.
It is their innate ability to recognize and distinguish [1] .
Giving the same capability to machine is vastly researched and debated area
under the computer vision domain which up to date has not found any complete
solution. Using a unique, measurable characteristic of a human beings known as
biometrics is always a good idea in this scenario. Biometrics can be classified
in to two sub domains, namely physiological biometrics and behavioral
biometrics. Iris-scan, face recognition, retina scan, fingerprint scan, and
palm scan can be given as examples for physiological biometrics while voice
scan, signature scan, and keystroke scan falls on to behavioral biometric sub
domain.
Among the biometrics fingerprint
scans are the most commonly used verification method. The solutions which uses
fingerprint scanners are of low budget when compared to the capital of other
solutions. Fingerprint is a very unique feature of a human where there is
almost zero percent chance of two people having the same thumbprint. Main
disadvantage of this method is user having to participate actively for the
verification process. Hardware used for iris scans, retina scans are costly
when compared with other solution which puts the main focus in to face
recognition. Face is an important part of who you are and how people identify
you, except for the occasion you being an identical twin. Humans use their
innate ability to identify and distinguish other humans, and without any
trouble humans are capable of doing this up to a satisfactory level.
Identification and verification are two important concepts which comes under face
recognition. Verification means where the system compares the given individual
with who that individual says they are, and gives a yes or no decision and
identification means the system comparing a given individual to all the other
individuals in a database and giving a ranked list of matches. Scientists began
to work on using the computer to recognize human faces since the mid-1960s.
There are many researches done in still image analytics and video stream
analytics [2]
up to date and facial recognition software’s have come a long way since then.
Most of the simple human face recognition systems does four major tasks, namely
image capturing, extraction, comparison and finally giving a yes or no answer for
the question or giving. There are many types of classifications but [3] gives a good
overview of above methods based on their approach of solving the face
recognition problem.
Knowledge-based Top-down
Model
The concepts of knowledge based top
down model is formed by converting the knowledge about human faces in to a set
of rules. So what forms a face, from the
very basic level the answer to that question would be two eyes, a nose and a
mouth. The relationship between the positions of eyes, mouth and nose in a
human face is taken into consideration. The important factor is the symmetry of
the placement of this components. But this is not always true, what happens if
a person’s face is deformed, will the same set of rules identify a face even if
it’s deformed. As said in the beginning of this document an additional
verification step should be done to avoid false matches (FAR).
The main problem in this approach is
deriving a set of rules from the information we have about a human face. The
strictness of the rule set will have a direct effect on the results of the
system. If the rules are too strict, some faces might not be identified as
faces increasing the false rejection rate of the approach and in contrast if
the rule set is not well formed or not strict enough it might identify
non-faces as faces. A knowledge based model can be explained using Yang and
Huang [4]
where they try to solve the problem by three levels of rules. The rules in the
higher level are more generalized as to what looks like a face and rules in the
ground level are established based on features of the face. This approach
failed to show considerable results and later Kotropoulous et al. [5] developed a method
which extends afore said by averaging intensities in each column of the image
and by finding abrupt changes in the intensities. The approach was successful
giving an 85.6% success rate.
Template Matching Approach
In this type of approach a standard
pattern of human face, most preferably a frontal image is stored first. The
correlation values with several types of standard patterns is calculated for nose,
mouth, eyes the face contours independently. Using Skin color information or
skin pixel information is one approach to do this. Yang Z. et al. [6] has conducted
experiments using this method on the FERET database and the study has shown
impressive results with future enhancements. Successful extraction of patterns
in a face can be difficult when it comes to real life scenario, the view of the
face might not be frontal always. If any match should be found through this
method, the template should be extracted from a same scale image which is not
possible in many scenarios. This is the main disadvantage of this method. Many
researchers [7] , [8] have tried to
minimize the dependency on scale by using sub templates and using deformable
objects.
There has been new development in
this area after the introduction of 3D template matching techniques which could
tolerate facial variations due to varying lighting conditions, different head
poses and facial expressions. General steps towards 3D face modeling and
matching consists of face model building using video or different still images
of a person and comparing the 3D model itself or a synthetic 2D image generated
using the system with a given image.
An approach
based on 3D morphable model was discussed in [9]
where 1200 real images of six people were tested. The system showed more than
90% accuracy surpassing global face recognition systems. A scheme based on the
analysis by synthesis framework was proposed by Lu et al. [10]
where a number of synthetic face images are generated with appearance
variations from the aligned 3D face model. These synthesized images were later
used to construct an affine subspace for each subject. Results were evaluated
using images taken of 10 subjects over a span of five weeks and it outperformed
the PCA based model (without synthesizing). Using the advantage of accumulation
of multiple frames in video Park and Jain [11] ,
proposed a way to compensate for low resolution, poor contrast and non-frontal
poses. The proposed scheme was tested on CMU’s Face in Action (FIA) video
database which consists of 221 subjects and 3D modelling of non-frontal images
showed 40% performance over the traditional non-frontal way.
Appearance based Methods
Neural Networks
Neural networks are systems of
program and data structures which approximates the operation of human brain.
This concept can be effectively used in to the domain of face recognition to
classify and identify faces through extensive training. Reason behind choosing
neural network approach to face recognition problem can be justified using the
ability of neural networks to remarkable derive meaning from complicated and
imprecise data. Neural networks have the capability of adaptive learning and
self-organizing which gives it edge over some other methods described in this
section.
In this approach the system is
trained to capture patterns in images of human face assuming that such is a
highly structured group of patterns which could be detected using well defined
boundaries. Training the system includes feeding it with face and non-face
patterns. The ability of neural networks to adaptively learn which means learn
to do a task by itself will take care of the problem from that point onwards.
The advantages is that there is no need to define the features of the human
face, the trained system will be capable of doing this by its own. Yet this
cannot be treated as a complete solution to the problem because of the
following shortcomings. There is no particular measure on up to what degree the
system should be trained and how many face and non-face images should be used.
Apart from that there is a high overhead on training the system since it takes
a lot of time and effort.
Perhaps the most significant work in
neural network based method was by Rowley, Baluja and Kanade [12] in which they have
proposed a method to detect faces which are invariant to the rotation up to any
degree. A bootstrap algorithm is for training the networks, which adds false
detections into the training set eliminating the difficult task of manually
selecting non-face training examples. This method was validated using three
sets of test data which consisted of entirely different images which were used
to train the system. The results observed were impressive and it revealed that
their method most heavily relied on the eyes, then on the nose, and then on the
mouth in human face.
Latha, Ganesan, Annadurai [13] has proposed a
method using Back Propagation of Neural Networks (BPNN) and Principal
Components Analysis (PCA) which was evaluated using 200 images from the Yale
database. A preprocessing step was carried out to normalize the image to
improve the robustness against illumination changes and occlusions. This method
has shown an accuracy of over 90%.
Support Vector Machines (SVM)
SVM is a learning technique which
was developed by V. Vapnik and his team at AT & T Bell Labs. It is new
paradigm for training polynomial, neural or Radial Based Functions (RBF) Classifier
[3] .
In this methods answers to linearly constrained quadratic problems is found
using a decomposition algorithms which guarantees the global optimality.
Optimizing equations of the
quadratic form can be overwhelming since the space is too dense and the memory
requirement tends to grow with the square of the number of data points. Osuna
et al. [14]
developed an algorithm which decomposes the problem to achieve global
optimality. They have successfully demonstrated the applicability of SVM to
face recognition by evaluating their concepts against a data set of 50,000 data
points. The two test sets used A, B consisted of 313 high quality and 23 mixed
quality images respectively and their results yielded at 97.1% and 74.2%.
SVM is not fully effective when
there are occlusions, which causes missing entries in feature vectors Hongjun
and Martinez [15] has proposed a
method by defining a criterion to minimize the probability of overlap. There is
no rule that only one method should be used to solve the problems in this
domain, Gadekar and Suresh [16]
have shown a collaborative approach of feature extraction, PCA and SVM which
provides an effective solution.
Eigenface Approach
Turk and Pentland [17]
demonstrated that significant improvements can be achieved by first mapping the
data into a lower dimensionality space this is known as classical eigenface
approach. Later a probabilistic visual learning model which is based on density
estimation of high dimensional space using an eigenspace decomposition was
developed by Moghaddam and Pentland [18] .
The latter showed better results when compared to the first approach, but there
method was only demonstrated using the images which were taken upright
(localized).
Statistical Approach
A probabilistic model for object
recognition using local appearances was proposed by Schneiderman and Kanade in [19] . This is
significantly different from the appearance based methods [12] , [13]
which model the object on full global appearance. In simple the latter methods
models the full face of the person. The statistical approach suggests that the
local appearances and local patterns are more unique when it comes to human
face. The intensity patterns around the human eyes are different from the
intensity patterns of the cheek, hence providing a suitable criteria to
identify a human uniquely. To represent this, statistics of local appearances
need to be modeled.
The Hidden Markov Model (HMM) is a
clustering algorithm based on high order statistics. Rajagopalan et al. [20] presents two schemes
in which the first scheme approximates the unknown distributions of the face
and the face-like manifolds wing higher order statistics (HOS). An HOS-based
data clustering algorithm is also proposed. In the second scheme, the face to
non-face and non-face to face transitions are learnt using a hidden Markov
model. The training set consisted of 2004 ”face” patterns, 4065 ”face-like”
patterns and 6364 additional ”non-face” patterns and HOS scheme showed slightly
better results than HMM scheme.
Independent Component Analysis (ICA)
can also be treated as a statistical model which reveals the underlying hidden
factors. The information describing a face may be contained in both linear as
well as high-order dependencies among the image pixels. These high-order
dependencies can be captured effectively by representation in ICA space. Linear
Discriminant Analysis (LDA) or Fisherfaces was another statistical approach
developed by Sir. R.A. Fisher in 1963. LDA is a robust mathematical model which
often produced good classification as same as some complex model. Many studies
have been done to compare the benefits of ICA and LDA. A good comparison of the
behavior of ICA and LDA over PCA method can be seen in the work of Delac et al.
[21] .
Further Lu and Plataniotis [22]
have identified a novel separability criterion which is called as Maximum
Separability Clusters (MSC) to prove that these methods can be used with large
data sets. Draper et al. [23]
conducted tests in view of finding the superior out of PCA and ICA methods. He used
two ICA architectures, two ICA algorithms, and three PCA distance measures only
to come to a conclusion that comparisons between PCA and ICA is complex.
Information Theory Approach
Kullback-Leibler divergence is symmetric
measure of the difference between two probability distributions H0 and H1.
This is applied in Information Theory approach of face recognition
where H1 denotes the event being the template is a face and H0
denoting template being a non-face. From the training sets Most Informative Pixels
(MIP) are extracted to improve the Kullback relative information [24] . A window is passed
over the image and distance from far space (DFFS) is calculated. If the DFSS to
the face cluster is lower than the DFSS to the non-face cluster it is assumed
that a face is within the image.
Bottom-up Feature Model
The researchers have been trying to
find feature invariant for the face detection. Robustness to poses, viewpoints
and changes in illumination had to be achieved for successful face detection.
They were inspired by the idea that humans are able to identify faces even when
there is large scale of pose variations, view point changes and changes in lightning.
Plus point was this method was the ability to detect faces even the poses were
different but it had its shortcomings where illumination played a big role. The
features of the face were badly affected by the false detection of edges due to
shadows etc. making the perceptual grouping algorithms useless.
In the early stage the work of Govindaraju et
al [25] , [26] suggested that a
face could be formed in terms of edges of the frontal image of the face. They
used the Marr-Hildreth operator to detect edges of the image which was later
processed and optimized using a thinning algorithm. The links between edges
were connected depending on the proximity and the orientation of the edges.
Even though it showed over 70% results on a 50 image data set, all the images
used were frontal making it less preferable to address a real life scenario.
A better method was proposed by Yow
and Cipolla [27] which consisted of two stages. Important
features called as “interest points” were extracted using the raw data from the
image as first step. At the final stage of processing the extracted interest
points were grouped together using Gestals principles. The points were labeled
using the knowledge gathered from the training data set. Each grouping was then
evaluated using Bayesian network. Results obtained by this method was in the
reach of 94%.
Using human skin color to
distinguish features of the human face was also a good idea. The studies showed
that difference of the appearance is due to the change of intensities rather
than the colors [3] .
Identifying skin like pixels in an image and grouping them together using
component analysis or clustering was done in this approaches. More recent studies in color based
segmentation [28]
and color invariant methodologies like Color SIFT [29] , [30]
has widen the use of such methods.
The turning point in human face
recognition was due to the introduction of a technique called Scale Invariant
Feature Transform or SIFT. Perhaps this is considered as the most influential
paper in computer vision [31] . Lowes work on this
area is considered as outstanding. He introduced a concept called scale
invariant features, the specialty of these features were they were invariant to
scale, translation, rotation and partial illumination. The so called features
or interesting points were so special that they could be identified even if the
image is scaled or rotated. SIFT algorithm consists of 4 phases namely, scale
space extrema detection, keypoint localization, orientation assignment and
building a key point descriptor. Each interest point will be described using
this feature vector. This method could identify faces even the faces were
partially occluded [32] . Lowe came up with a
more efficient method to optimize the search time using nearest neighbors of
points which was far more efficient than exhaustive searches [33] .
This method even tolerates affine transformation up to some degree, but when it
comes to face detection and face recognition more tolerance on affine
transformation is not needed.
An efficient method for face
recognition and retrieval using Local Binary Patterns (LBP) and SIFT was
proposed by Tayade and Bansode [34] .
Their system took an image as input, filtered it and represented in sparse
matrix deriving SIFT and LBP features. After that a transition matrix was
computed using the inner distances and inner context to retrieve the results.
This system identified images at an accuracy of 81.25 %.
Mohamed Aly compared the
performances of Eigenfaces, Fisher Faces and SIFT algorithms using Nearest
Neighbor Algorithm on two standard set of databases, The AT & T database
and the Yale database. His results have clearly shown the superiority of SIFT
over the other algorithms [35] .
Amit Kr. [36]
came up with an improved solution which uses SIFT and Knearest neighbor
classifier in which he compares the recognition rates of five different face
recognition algorithms including SIFT. His method tops the table with a success
rate of 97.1% where SIFT alone has only achieved 96.3%. Results were also
verified using standard ORL database. There are many improvements done to the
SIFT algorithms and there are versions such as Fast Approximated SIFT and Very
Fast SIFT. Grabner [37] has done a
comparison between SIFT and a more developed version of SIFT named fast
approximated SIFT in terms of computational cost and speed. Study by Alwarin [38] demonstrates another
method called VF-SIFT. He proves that by using a suitable tradeoff on up to
what degree the features are matched will result in significant decrease of
computational time, sometimes giving 1250 times speed over the normal method. Both
of this results sets gives a hint to us that further improvement of this method
will be highly likely to in future.
Speed Up Robust Features (SURF) is
also method which uses the power of feature descriptors [39] .
It was a novel scale and rotation-invariant detector and descriptor. The main
advantage of SURF is its speed, it approximates or even outperforms previously
proposed schemes with respect to repeatability, distinctiveness, and robustness
which was a progress. The reduction of the computational time allowed SURF to
match features across bigger data sets. This improvement was gained through the
use of concepts like integral images and interpolation. A simple Hessian matrix
was used for the approximation which uses box type convolution filters
initially proposed by Viola and Jones [40] .
These box type convolution filters are known as Haar cascades or wavelets. SURF
has a 64 dimension feature vector which gives the same results as SIFT which
uses 128 dimension vector but with a reduced cost of computation. The accuracy
of the hessian detector for the application of camera self-calibration and 3D
reconstruction is also discussed in the same article.
Panchal [41] has compared these
two methods in terms of speed and the number of interest points detected. Tests
was carried out using two images where SIFT detected 892 and 934 interest
points while SURF only detected 281 and 245. But the number of feature points
matched were 41 and 28, resulting in 1.543 s and 0.546 s computational time.
The summary was that even though SIFT detected much more interest points it was
not really needed. The results were not verified using a standard data set
which included more complex scenes to come to a conclusion on what method is
better over the other.
In his Masters thesis Guerro [42] compared the
performance of three different algorithms namely SIFT, SURF and Fast [43] . The results confirmed the above mentioned
observation where SIFT feature detection was more far ahead of SURF but it was
less effective since SURF did the same matches with less number of matched
points. The study was carried in civil engineering background where the most of
the images contained more edges. The images which had more textures were easily
detected by SIFT and SURF. He suggested that FAST corner detector is a better
detector when compared to SIFTs Difference of Gaussian (DOG) and SURFs box
filter based methods.
A Comparison of SIFT, PCA-SIFT and
SURF by Jaun and Gwon [44]
using K-Nearest Neighbor (KNN) and Random Sample Consensus (RANSAC) has showed
that SIFT presents stability in most of the situations but was slow when it
comes to computing. SURF was the fastest one with good performance as same as
SIFT. PCA-SIFT showed its advantages in rotation and illumination changes. The
work was impressive since it has evaluated the performances of these algorithms
against aspects such as time, scale, rotation, blur, illumination and affine
transformation, and results are as tabulated below.
Algorithm
|
Time
|
Scale
|
Rotation
|
Blur
|
Illumination
|
Affine
|
SIFT
|
Common
|
Best
|
Best
|
Best
|
Common
|
Good
|
PCA-SIFT
|
Good
|
Common
|
Good
|
Common
|
Good
|
Good
|
SURF
|
Best
|
Good
|
Common
|
Good
|
Best
|
Good
|
A
Comparative Study of SIFT and its Variants by Wu et al. [45]
compares the relative performances of SIFT, PCA-SIFT, CSIFT, GSIFT [46] , ASIFT [47] and SURF over
factors such as scale, rotation, illumination, blur, affine transformation and
time cost. The results obtained are tabulated below to get a comprehensive idea
about how these algorithms perform under different circumstances.
Algorithm
|
Scale & Rotation
|
Illumination
|
Blur
|
Affine Transformation
|
Time Cost
|
SIFT
|
Best
|
Good
|
Better
|
Good
|
Better
|
PCA-SIFT
|
Better
|
Better
|
Better
|
Good
|
Better
|
GSIFT
|
Good
|
Best
|
Best
|
Good
|
Better
|
CSIFT
|
Best
|
Better
|
Good
|
Better
|
Good
|
SURF
|
Common
|
Common
|
Common
|
Common
|
Best
|
ASIFT
|
Good
|
Common
|
Common
|
Best
|
Common
|
Table 2: Comparison of SIFT and it’s Variants
References
[1]
|
R. S. Sandhu and P. Samarathi, "Access Control:
Principles and Practice," IEEE Communication Magazine, pp. 40-28,
1994.
|
[2]
|
W. Zhao, R. Chellappa, A. J. Phillips and A.
Rosenfield, "Face Recognition: A Literature Survey," ACM
Computer Journal, vol. 35, no. 4, pp. 399-458, 2003.
|
[3]
|
M. Yang, A. Ahuja and D. Kriegman, "A Survey on
Face Detection Methods," 1999.
|
[4]
|
G. Yang and G. S. Huang, "Face Recognition in
Complex Backgrounds," Pattern Recognition, vol. 27, no. 1, pp.
53-63, 1994.
|
[5]
|
C. Kotropoulos, A. Tefas and I. Pitas, "Frontal
Face Authentication using Morphological Elastic Graph Matching," Image
Processing, IEEE Transactions, vol. 9, no. 4, pp. 555-560, 2002.
|
[6]
|
Z. Jin, Z. Lou, J. Yang and Q. Sun, "Face
detection using template matching and skin-color information," in International
Conference on Intelligent Computing, Hefei, 2005.
|
[7]
|
T. Sakai, M. Nagao and S. Fujibayashi, "Line
extraction and pattern detection in a photograph," The Journal of the
Pattern Recognition Society, vol. 1, no. 3, pp. 233-236, 1969.
|
[8]
|
A. Yuille, P. Halinan and D. Cohen, "Feature
Extraction from Faces using Deformable Templates," International
Juornal of Computer Vision, vol. 8, no. 2, pp. 99-111, 1992.
|
[9]
|
J. Huang, B. Heisele and V. Blanz, "Component-Based
Face Recognition with 3D Morphable Models," Springer, Berlin, 2003.
|
[10]
|
X. Lu, H. Rein-Lien, A. K. Jain, B. K. P. and B. K.
P., " Face recognition with 3D model-based synthesis," in International
Conference on Biometric Authentication volume 3072 of Lecture Notes in
Computer Science, Verlag, 2004.
|
[11]
|
U. Park and A. K. Jain, "3D Model-Based Face
Recognition in Video," Proc. International Conference on Biometrics,,
pp. 1085-1094, 2007.
|
[12]
|
H. A. Rowley, S. Baluja and T. Kanade, "Rotation
Invariant Neural Network-based Face Detection," in Computer Vision
and Pattern Recognition, 1998. Proceedings. 1998 IEEE Computer Society
Conference on, Santa Barbara, CA, 1998.
|
[13]
|
P. Latha, L. Ganesan and S. Annadurai, "Face
Recognition using Neural Networks," Signal Processing International
Journal, vol. 3, no. 5, pp. 153-160, 2009.
|
[14]
|
A. Osuna, R. Freund and F. Girosi, "Training
Support Vector Machines: an Application to Face Detection," in Conference
on Computer Vision and Pattern Recognition, Puerto Rico, 1997.
|
[15]
|
H. Jia and A. M. Martinez, "Support Vector
MAchnes in Face Recognition with Occlusions," The Department of
Electrical and Computer Engineering, Ohio State University, Columbus, OH.
|
[16]
|
A. D. Gadekar and S. S. Suresh, "Face
Recognition Using SIFT-PCA Feature Extraction and SVM Classifier," IOSR
Journal of VLSI and Signal Processing, vol. 5, no. 2, pp. 31-35, 2015.
|
[17]
|
M. Turk and A. Pentland, "Eigenfaces for Face
Recognition," Journal of Cognitive Neuroscience, vol. 3, no. 1,
pp. 71-86, 1991.
|
[18]
|
B. Moghadam and A. Pentland, "Probabilistic
Visual Learning for Object Detection," in International Conference on
Computer Vision, Cambridge, MA, 1995.
|
[19]
|
H. Schneiderman and T. Kanade, "Object
Detection Using the Statistics of Parts," International Journal of
Computer Vision, vol. 56, no. 3, pp. 151-177, 2004.
|
[20]
|
A. N. Rajgopalan, K. S. Kumar, J. Karlekar, R.
Manivasakan, M. M. Patil, U. B. Desai and P. G. C. S. Poonacha, "Finding
faces in photographs," in Sixth Annual Conference on Computer Vision,
Mumbai, 1998.
|
[21]
|
K. Delac, G. Grgic and S. Grgic, "Independant
Comaparative Study of PCA, ICA, and LDA on the FERET Data Set," Wiley
Periodicals, Inc., London, 2006.
|
[22]
|
J. Lu and K. N. Plataniotis, "Boosting Face
Recognition on a Large Scale Database," in International Conference
on Image Processing, Rochester, 2002.
|
[23]
|
A. Bruce, A. Draper, K. Baek, M. S. Bartlett and J.
R. Beveridge, "Recognizing faces with PCA and ICA," Computer
Vision and Image Understanding, vol. 91, pp. 115-137, 2003.
|
[24]
|
A. Pentland, B. Moghaddam and T. Starner,
"View-based and modular eigenspaces for face recognition," in IEEE
Conference on Computer Vision & Pattern Recognition, Seattle, WA,
1994.
|
[25]
|
V. Govindaraju, "Locating Human faces in
photographs," International Journal of Computer Vision, vol. 19,
no. 2, pp. 126-146, 1996.
|
[26]
|
V. Govindaraju, D. B. Sher, R. K. Srihari and S.
Srihari, "Locating human faces in newspaper photographs," in Computer
Vision and Pattern Recognition, 1989. Proceedings CVPR '89., IEEE Computer
Society Conference, San Diego,CA, 1989.
|
[27]
|
K. C. Yow and R. Cipolla, "Feature-Based Human
Face Detection," Image and Vision Computing, vol. 15, no. 9, pp.
715-735, 1997.
|
[28]
|
Y. Tayal, R. Lamba and S. Padhee, "Automatic
Face Detection using Color Based Segmentation," International Journal
of Scientific and Research Publications, vol. 2, no. 6, 2012.
|
[29]
|
A. E. Abdel-Hakeem and A. A. Farag, "CSIFT: A
SIFT Descriptor with Color Invarint Characteristics," in IEEE
Computer Society Conference on Computer Vision and Pattern recognition,
New York, 2006.
|
[30]
|
A. Singh, S. K. Singh and S. Tiwari,
"Comparison of Face Recognition Algorithms on Dummy Faces," The
International Journal of Multimedia & Its Application (IJMA), vol. 4,
no. 4, pp. 121-135, 2012.
|
[31]
|
D. Lowe, "Object Recognition from Local
Scale-Invariant Features," in International Conference on Computer
Vision, Corfu, 1999.
|
[32]
|
Y. Cai, "Invariant Local Features for Face
Detection," Submitted for Publication, Vancouver, 2008.
|
[33]
|
D. Lowe, "Distinctive Image Features from
Scale-Invariant Keypoints," International Journal of Computer Vision,
vol. 60, no. 2, pp. 91-110, 2004.
|
[34]
|
S. R. Tayade and S. Bansode, "An Efficient Face
Recognition and Retrieval Using LBP and SIFT," International Journal
of Advanced Research in Computer and Communication Engineering, vol. 2,
no. 4, pp. 1769-1773, 2013.
|
[35]
|
A. Mohammed, "Face Recognition using SIFT
Features," California Institute of Technology, California, 2006.
|
[36]
|
G. Amith Kr and Twisha, "Improved Face
Recognition Technique using SIFT," in International Conference on
Advances in Engineering & Technology, Chandigarh, 2014.
|
[37]
|
M. Grabner, H. Grabner and H. Bischof, "Fast
Approximated SIFT," in Asian Conference on Computer Vision,
Hydrabad, 2006.
|
[38]
|
F. Alhwarin, D. Ristic-Durrant and A. Graser,
"VF-SIFT: Very Fast SIFT Feature Matching," Springer-Verlag,
Berlin, 2010.
|
[39]
|
H. Bay, A. Ess, T. Tuytelaars and L. Van Gool,
"Speed -Up Robust Features (SURF)," Elsevier, Amsterdam, 2008.
|
[40]
|
P. Viola and M. J. Jones, "Robust Real-Time
Face Detection," International Journal of Computer Vision, vol.
57, no. 2, pp. 137-154, 2004.
|
[41]
|
P. M. Panchal, S. R. Panchal and S. K. Shah, "A
Comparison of SIFT and SURF.," International Journal of Innovative
Research in Computer and Communication Engineering, vol. 1, no. 2, pp.
323-327, 2013.
|
[42]
|
M. Guerrero, "A Comparative Study of Three
Image Matcing Algorithms: Sift, Surf, and Fast," Utah State University,
Logan, Utah, 2011.
|
[43]
|
M. Trajkovic and M. Hedley, "Fast Corner
Detection," Image and Vision Computing, vol. 16, pp. 75-87, 1998.
|
[44]
|
L. Juan and O. Gwon, "A Comparison of SIFT,
PCA-SIFT and SURF," International Journal of Image Processing (IJIP),
vol. 3, no. 4, pp. 143-152, 2009.
|
[45]
|
J. Wu, Z. Cui, V. S. Zheng, P. Shao, D. Su and S.
Gong, "A Comparative Study of SIFT and its Variants," Measurement
Science Review, vol. 13, no. 3, pp. 122-131, 2013.
|
[46]
|
E. N. Mortensen, H. Deng and L. Shapiro, "A
SIFT descriptor With Global Context," in IEEE Computer Society
Conference on Computer Vision and Pattern Recognition , San Diego, CA,
2005.
|
[47]
|
J. M. Morel and G. Yu, "ASIFT: A New Framework
for Fully Affine Invariant Image Comparison," SIAM Journal on Imaging
Sciences, vol. 2, no. 2, pp. 438-469, 2009.
|
Subscribe to:
Posts (Atom)