원문정보
초록
영어
So far, several methods have been proposed for gender estimation based on training processes, using several images per person. In this paper, we introduce a new gender estimation method based on single image per person. First, reference images, called general faces, are constructed by averaging male and female images in the dataset respectively. Then, important face features such as eyes and mouth are highlighted by Wavelet fusion of Sobel Detector, and Laplacian of Gaussian images. The highlighted image is segmented into seven partitions. In the obtained image, those segments like eyes and eyebrows with more variations (containing more information) play more significant role in the gender estimation process. Weight of each segment is proportional to its entropy. Finally, DCT coefficients of each segment in the input facial image are compared to their counterparts in general faces. Considering weight of each segment and fuzzy rule sets, the input image is assigned to the closest class. The proposed method is evaluated on FERET and AR databases. Experimental results show that the proposed method is tolerant of variations such as facial expression, occlusion, and rotation.
목차
1. Introduction
2. Related Work
3. The Proposed Gender Estimation Algorithm
3.1. Image Datasets
3.2. General Face Construction
3.3. Image Partitioning
3.4. Face Features Highlighting
3.5. Image Entropy
3.6. Discrete Cosine Transform
3.7. Distance Measures
3.8. Fuzzy System
4. Experimental Results
4.1. Distance Measures Study
4.2. Number of Images Effect
4.3. Fuzzy Logic Effect
4.4. Performance Evaluation Using Different Datasets
4.5. Performance Evaluation Using Non-frontal Facial Images
4.6. Performance Evaluation with Occulted Images
4.7. Performance Evaluation Using an Ethnic Dataset
4.8. Comparison with Related Works
5. Conclusion and Future Work
Acknowledgements
References