Face Recognition Using Discrete Cosine Transform for Global and Local Features

Computer Science – Computer Vision and Pattern Recognition

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

face recognition; biometrics; person identification; authentication; discrete cosine transform; DCT; global local features; Pr

Scientific paper

Face Recognition using Discrete Cosine Transform (DCT) for Local and Global Features involves recognizing the corresponding face image from the database. The face image obtained from the user is cropped such that only the frontal face image is extracted, eliminating the background. The image is restricted to a size of 128 x 128 pixels. All images in the database are gray level images. DCT is applied to the entire image. This gives DCT coefficients, which are global features. Local features such as eyes, nose and mouth are also extracted and DCT is applied to these features. Depending upon the recognition rate obtained for each feature, they are given weightage and then combined. Both local and global features are used for comparison. By comparing the ranks for global and local features, the false acceptance rate for DCT can be minimized.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Face Recognition Using Discrete Cosine Transform for Global and Local Features does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Face Recognition Using Discrete Cosine Transform for Global and Local Features, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Face Recognition Using Discrete Cosine Transform for Global and Local Features will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-704734

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.