Next Article in Journal
A New Void Fraction Measurement Method for Gas-Liquid Two-Phase Flow in Small Channels
Previous Article in Journal
Laterally Driven Resonant Pressure Sensor with Etched Silicon Dual Diaphragms and Combined Beams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Body-Based Gender Recognition Using Images from Visible and Thermal Cameras

Division of Electronics and Electrical Engineering, Dongguk University, 30, Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea
*
Author to whom correspondence should be addressed.
Submission received: 30 December 2015 / Accepted: 21 January 2016 / Published: 27 January 2016
(This article belongs to the Section Physical Sensors)

Abstract

:
Gender information has many useful applications in computer vision systems, such as surveillance systems, counting the number of males and females in a shopping mall, accessing control systems in restricted areas, or any human-computer interaction system. In most previous studies, researchers attempted to recognize gender by using visible light images of the human face or body. However, shadow, illumination, and time of day greatly affect the performance of these methods. To overcome this problem, we propose a new gender recognition method based on the combination of visible light and thermal camera images of the human body. Experimental results, through various kinds of feature extraction and fusion methods, show that our approach is efficient for gender recognition through a comparison of recognition rates with conventional systems.

1. Introduction

With the development of digital smart systems, such as smart surveillance, or human-computer interaction systems, there are new demands on digital imaging. Many different kinds of information must be obtained in order to perform different system tasks. Information on gender is useful to these smart systems [1,2,3]. For example, the human-computer interaction systems should know gender information in order to respond to users appropriately. The system should respond Mr. or Ms. to users based on gender, and could show specific information to either male or female users. In some other human-computer interaction systems, the systems may need to recognize the individual users and/or their facial expressions in order to interact with them. In this case, gender information can serve as a soft biometrics feature in order to assist recognition results [1,2]. In a surveillance system, knowing the gender information could help the system restrict users from specific restricted areas (the areas are only used by male or female) [3]. The gender information could also be used in demographic collection [1]. Using gender information, the surveillance systems can count the number of males and females who enter a shopping malls or stores. That information is useful for collecting demographic data for commercial purposes [1,2,3].
Because of the many useful applications of gender information, many previous studies have proposed methods for gathering gender information. These previous methods can be grouped into two main categories, the voice-based recognition group [4,5] and image-based recognition group [6,7,8,9,10,11,12,13,14]. In voice-based recognition methods, Mel-frequency cepstral coefficient (MFCC) and pitch values of the human voice are frequently used for gender recognition [4]. High recognition accuracy was obtained by combining the two methods [4]. Although this approach produces high recognition accuracy, noise and microphone quality greatly affects the results.
Gender can also be recognized by using computer vision methods. In this type of method, parts of the human body that contain gender information are captured using camera sensors and then used for gender recognition. In previous researches, images of fingerprints, faces, and/or the entire body were used. In the research of Arun et al. [6], fingerprint images are used, and the feature vector describing the ridge thickness to valley thickness ratio (RTVTR) and the ridge density values is obtained. After that, they used the support vector machine (SVM) to classify subjects into male and female groups. Along with fingerprint analysis, computer vision researchers have analyzed gender using facial images. Visually, it is easy to discriminate gender based on the face. Because of this, many previous studies have captured facial images and used them for the gender recognition problem [7,8,9]. In order to enhance recognition accuracy, fingerprint and facial images are combined to solve the gender recognition problem [10,11]. Although the systems using fingerprint, facial, and combined fingerprint-facial images have been proven quite effective for gender recognition problems, they are all limited in that these systems require the cooperation of users when capturing images. Without the cooperation of users, the quality of fingerprint and face images is not sufficient for gender recognition. In addition, in order to capture fingerprint and face images, the system requires a very small distance between the camera sensors and users. This limitation prevents the recognition systems from being applied in surveillance systems.
To overcome the limitations of fingerprint and face-based gender recognition systems, a new biometric modality based on the full body is used. For this reason, the recognition system captures images of the entire body instead of fingerprint and/or face. This approach limits the difficulties presented by low quality of capture images and the cooperation requirement of the fingerprint-based and face-based systems. In the study by Cao et al. [12], researchers applied the histogram of oriented gradient (HoG) feature extraction method on body images to extract the image feature. After that, they use the classification method for gender classification. Later, the research by Guo et al. [3] enhanced recognition performance by using the biologically-inspired feature extraction method (BIF) to extract image features and used linear support vector machine (SVM) for classification the gender. Through experiments, these studies show that the full body is sufficient for gender recognition up to an accuracy of an about 80% in classification rate using a public database (MIT database). The main limitation of these studies is that they used only visible light images for the recognition problem. The body images have large variation in pose and image texture due to the different clothing and hair styles, accessories etc. Consequently, these differences affect the extracted feature and recognition performance. In addition, this method fails if the images are too light or too dark. In that case, the detection of body is firstly failed that causes the wrong gender recognition result. Some other methods use 3-D body-shape for gender recognition [13,14], but in order to obtain the 3-D body-shape, a laser scanner is necessary, and this requires user cooperation, an expensive capturing device, and time for processing. Therefore, this method is hard to apply broadly in surveillance systems since these may require low cost and without cooperation of users. There is no previous research that uses the combination of visible light and thermal images of the full human body for the gender recognition problem. In order to overcome the problem of previous researches on the gender recognition problem, we propose a new gender recognition method that is based on the combination of visible light and thermal images of body. By using the appearance of the full body in both visible light and thermal images, our proposed method captures richer gender information from body for recognition purposes than previous researches in [3,12]. Therefore, the recognition accuracy is enhanced compared to the previous methods that use only visible light images for recognition problem. The research is novel in the following four ways compared to previous studies:
  • This is the first study using both the visible light and thermal images of the full human body for gender recognition.
  • Based on the detected boxes of body in both visible and thermal images, the features for gender recognition are extracted from the visible light and thermal image. This is accomplished by using the histogram of oriented gradient (HoG) method with principal component analysis (PCA) in order to reduce the feature dimension, processing time, and the effects of noise on the extracted features.
  • Gender classification using the features from visible light and thermal images are made by using two different SVMs classifiers.
  • A score level fusion is performed to combine the two classification scores by the two SVMs using another SVM classifier in order to recognize the gender of human.
In Table 1, we summarize the previous gender recognition methods and their strengths and weaknesses in comparison to our proposed method. The remainder of this paper is structured as follows: In Section 2, we describe the proposed gender recognition method using the combination of visible light and thermal images of the full human body. In Section 3, we present experimental results using our proposed method applied in a surveillance system. Finally, the conclusions are shown in Section 4.
Table 1. Summary of previous studies on image-based gender recognition.
Table 1. Summary of previous studies on image-based gender recognition.
CategoryMethodStrengthWeakness
Fingerprint-based gender recognitionUsing fingerprint image for gender recognition [6,10,11].High accuracyRequires the cooperation of users.
The accuracy is affected by quality and the resolution of fingerprint image.
Face-based gender recognitionUsing human face for gender recognition [7,8,9,10,11].High accuracyRequires the cooperation of users.
It’s very hard to recognize gender for very young people.
Body-based gender recognitionUsing only visible images of human body for gender recognition [3,12], with 3D shape model [13,14].Does not require the cooperation of users.Recognition accuracy is strongly affected by illumination conditions, body poses, the random appearance of image texture on body region such as clothes, accessories etc.
Recognition accuracy is lower than using face-based gender recognition approach.
Combining the visible and thermal images of human body using score level fusion using SVM for gender recognition (Proposed Method)Does not require the cooperation of users.
Enhances the recognition result compared to the systems that use only visible images for gender recognition.
Reduces the affects by illumination condition, the body poses, the random appearance of image texture on boy region such as clothes, accessories etc.
Requires longer processing time than singly visible images.
Recognition accuracy is still lower than using face-based gender recognition approach.

2. Proposed Method

2.1. Overview of the Proposed Method

In previous studies, it has been shown that images of the human body contain gender information [3,12,13,14]. Based on this information, body images were used for the gender recognition problem. The overall procedure of our proposed method is depicted in Figure 1.
Figure 1. Overall procedure for the proposed method.
Figure 1. Overall procedure for the proposed method.
Sensors 16 00156 g001
As shown in the figure above, our proposed method uses the appearance of the human body in two different kinds of images, a visible light image and a thermal image for gender recognition purposes. Using visible light and thermal cameras, we capture images that contain corresponding body regions of the same person. The images are first preprocessed using the human detection method [15]. In the second step, our method extracts the image’s features using feature extraction methods such as HoG or multi-level local binary pattern (MLBP). The details of human region detection and image feature extraction will be explained in Section 2.2. After that, the extracted features from the two types of images are combined to construct the final feature for gender recognition. Finally, the gender of users is recognized using SVM.
In order to combine the recognition results of visible and thermal images, our proposed method performs two different combination approaches, feature and score level fusion. The overview of these combination approaches is shown in Figure 2 and Figure 3 for feature level fusion and score level fusion, respectively. As shown in Figure 2, the feature level fusion approach is formed by combining the extracted feature vector of visible and thermal images together. In our research, the combination is performed by concatenating the extracted features of visible and thermal images. Because the feature vectors of the two images are concatenated, the combined feature vector contains the gender information for the two types of images. However, this process causes an increase in feature dimension and associated noise. In order to reduce the feature’s dimension, our proposed method performs a subspace method, based on PCA, on the combined feature [16]. Consequently, we obtain the new feature vector in a lower dimension space than original feature vector. By using PCA, we not only reduce the dimension of the feature vector but also reduce the effects of noise and the processing time of the recognition system. Finally, with the extracted feature (after PCA), we recognize the gender using SVM.
In the second combination approach, we combine the visible and thermal images of the full body using score level fusion as shown in Figure 3. To accomplish this, we first perform gender recognition of visible light and thermal images separately using the first layer of SVM classification.
Figure 2. Feature level fusion approach for combining visible light and thermal images.
Figure 2. Feature level fusion approach for combining visible light and thermal images.
Sensors 16 00156 g002
Figure 3. Score level fusion approach for combining the visible light and thermal images.
Figure 3. Score level fusion approach for combining the visible light and thermal images.
Sensors 16 00156 g003
This layer consists of two SVM classifiers for visible light and thermal images, respectively. For the score level combination, we obtain the decision values of classifiers instead of the predicted class label. The decision values represent the probability that an image belongs to either the male or female class. For example, we assign the male and female classes labels of 1 and −1, respectively. Then, if the decision value is closed to 1, we can think that the image is male. If the decision value is closed to −1, we can think that the image is female. The decision values stand for the probability that an image belongs to one of two classes. Using these first two SVM classifiers, we can obtain two score values (the two decision values). In final step of score level fusion approach, our proposed method uses a second SVM classification layer to recognize the gender using the two input scores from the first SVM classification layer. In our experiments, we will evaluate the recognition performance of a single kind of image (either visible or thermal image) and the two combination methods.

2.2. Image Acquisition and Human Detection from Image Sequences

In order to recognize gender using body images, we propose capturing visible light and thermal images using a dual visible-thermal camera. A preprocessing step is applied to these images in order to detect the human region and locate the position of the body in these captured images. In our experiment, because there is no public surveillance database that contains both visible light and thermal images, we collect our own database for experiments using our lab-made devices. The structure of our lab-made dual visible-thermal camera is shown in Figure 4. The capturing device in our research consists of two cameras, a visible light camera and a far-infrared light (FIR) camera. We use a low-cost web-camera for the visible light camera since our method is designed for surveillance systems [17]. This camera can capture images with an image resolution of 2 mega-pixels at 30 frames per second (fps). In order to capture the thermal images, either a near-infrared light (NIR) or far-infrared light (FIR) camera can be used. In our research, we use the FIR camera because this kind of camera is more suitable for surveillance systems than NIR camera. The NIR camera can captures signal with wavelengths between 0.75 μm and about 1.4 μm and normally requires an additional NIR illuminators to capture thermal images. This requirement makes the NIR camera ineffective for surveillance systems that need to monitor objects in a far distance, and in an uncontrolled working environment. In contrast to the NIR camera, the FIR camera captures image signal with wavelength range of 8–12 μm and it does not require additional illuminators for capturing images. Therefore, the FIR camera is more appropriate for surveillance applications than the NIR camera. In our lab-made devices, we use the Tau2 [18] commercial thermal camera. In order to combine the two cameras, we attached them on a panel as shown in Figure 4 to fix the position of the cameras in the horizontal direction and reduce the disparity between the two cameras. In Figure 5a,b we show some examples of images captured by our lab-made dual visible-thermal camera.
Figure 4. Structure of the dual visible-thermal camera used in our research.
Figure 4. Structure of the dual visible-thermal camera used in our research.
Sensors 16 00156 g004
Figure 5. Example of visible and thermal images captured by our lab-made dual visible-thermal camera, and the results of human detection using Lee et al.’s method: (a) a visible light image; (b) a thermal image captured at the same time and same scene with the visible light image; the detection results with (c) the visible light image; and (d) the thermal image.
Figure 5. Example of visible and thermal images captured by our lab-made dual visible-thermal camera, and the results of human detection using Lee et al.’s method: (a) a visible light image; (b) a thermal image captured at the same time and same scene with the visible light image; the detection results with (c) the visible light image; and (d) the thermal image.
Sensors 16 00156 g005
With the captured visible and thermal images, our proposed method performs a preprocessing step to extract the body images in both images. To do this, we use the method proposed by Lee et al. [15] for human detection. Using this method on both images, we enhance the detection accuracy when compared to using a single type of image. In Figure 5c,d, we show some examples of detection results using Lee et al.’s method [15]. As shown in these figures, the region of interest (ROI) of the human body is well located in the captured visible and thermal images. Using this result, we extract the body images and use them for gender recognition in the later steps of our proposed method.

2.3. Feature Extraction Methods for Gender Recognition

After identifying body boxes within images, the next step in our proposed method is to extract image features from the images. The simplest way to use these images is to use the raw pixels directly. However, direct use of the raw pixels is not a good choice due to the large variation in these images such as clothing color, illumination change, and hairstyle. In order to extract the more distinct gender features for gender recognition purpose, we use two different feature extraction methods. We use the HoG and MLBP methods to evaluate the performance of recognition system for inter-experimental comparison.

2.3.1. Histogram of Oriented Gradient

The HoG feature extraction technique was successfully applied on the human detection problem [19] and the gender recognition problem [12]. In previous research on gender recognition by Cao et al. [12], the HoG technique was firstly used on the problem and had good performance. As indicated by its name, the HoG feature extraction method extracts image features by collecting the magnitude-weight and direction of edge features in sub-blocks of an image. Consequently, we can obtain a map of age of the entire image. For our problem, the gender information is mainly measured based on body-shape and the different kinds of edge in a body image. Therefore, the HoG feature technique is efficient for gender recognition as indicated in [12]. In our studies, we use the HoG technique for extracting image features for the gender recognition problem.

2.3.2. Multi-Level Local Binary Pattern

As mentioned in previous sections, the human body images-based gender recognition problem is complicated by illumination variations in the captured body images. Previous studies in computer vision have proven that the local binary pattern (LBP) is a powerful image feature extraction method. The main advantage of the LBP is that it is invariant to the illumination levels of an image. Based on this characteristic, the LBP has been successfully used for many biometrics systems such as finger-vein recognition [20], face recognition [21], age estimation [16,22,23], gender recognition [24], and face re-identification [25]. In previous research on human age estimation, the author proposed the use of a multi-level LBP methodology instead of simple LBP for age estimation problem [16,22]. By accumulating the histogram of uniform and non-uniform LBP features in an image, we can identify features such as lines (edges), corners, spots etc. This methodology yielded good results for age prediction. In addition, the MLBP feature (image feature extracted by MLBP method) has been proven to outperform the LBP feature [16,22]. Inspired from this research, we use MLBP feature extraction method for the gender recognition problem. The LBP operator is mathematically defined in Equation (1):
L B P P , R =   p = 0 P 1 s ( x c x p ) × 2 p , where s ( x ) = { 0   i f   x < 0 1   o t h e r w i s e
In Equation (1), R value indicates the radius of a circle that the surrounding pixels are located within, and P is the number of surrounding pixels. As shown in this equation, we see that an illumination invariant characteristic is obtained using the LBP method. By comparing the center pixels with its P surrounding pixels, the LBP feature extraction method works as an adaptive thresholding method for extracting the image texture feature. Consequently, a LBP code of P bits is obtained to encode each pixel in an image. Although the illumination conditions could be changed by changing image-capturing conditions, the LBP method offers the same image features regardless of illumination. By using various values of R and P, we can extract the image feature at different scales and resolutions. In order to construct the image feature, the extracted LBP codes are first divided into uniform and non-uniform codes [16,22]. The uniform codes are defined as containing at most two bit-wise transitions from 0 to 1 (or 1 to 0). The other kinds of codes are defined as non-uniform code. For example, 00000000 and 011111000 are uniform codes because they contain two bit-wise transitions from 0 to 1 (or 1 to 0), while 01010100 and 01100110 are non-uniform codes because they contains six and four bit-wise transitions, respectively. The uniform codes contain useful texture information such as edge, corner, spot etc. On the other hand, the non-uniform code describes very complex image texture features. These complex textures could also be caused by noise. Therefore, this kind of texture feature is not sufficient for making an image feature. Similar to the age estimation problem, we see that the direction of the texture (edge, corner etc.) is not as important for the gender recognition problem as the appearance and the number of texture feature in image. Therefore, we further group the uniform LBP codes by considering patterns that have similar texture shape but could be going in different directions. By performing this step, we make the rotation invariant LBP code [16,22]. Based on that, we assign each group of uniform rotation invariant texture patterns a specific decimal code from 0 to P, while all the non-uniform patterns are grouped together and assign a decimal code of (P + 1). With the assigned decimal codes, we can easily create a histogram of the appearance of texture features and use them for the gender recognition problem.
In order to extract image features more efficiently, previous researchers [16] divided the image into sub-blocks and constructed the image feature by concatenating the extracted features of each sub-block. This scheme was successfully applied for face-based human age estimation [16,22] and the extracted feature is called a single level LBP feature. By using this method, the extracted image features not only contain the global image feature but also the local image feature. As reported in previous research, the MLBP feature outperforms the single level LBP feature. In Figure 6, we demonstrate methodology of MLBP feature extraction method. In this figure, M and N are the number of sub-blocks in the vertical and horizontal directions. By combining several single level LBP features, we form the MLBP feature for the gender recognition problem.
Figure 6. Method for extracting the MLBP feature from an image.
Figure 6. Method for extracting the MLBP feature from an image.
Sensors 16 00156 g006

2.3.3. Feature Combination and Gender Recognition Using SVM

The final steps of our proposed method are the combination of features from the visible light and thermal images and the recognition of gender. As shown in Section 2.1 and Figure 2 and Figure 3, the combination of visible and thermal images can be performed using either the feature level fusion or score level fusion approach. In feature level fusion, the two extracted features are concatenated together to form the final feature for gender recognition as shown in Equation (2)
f = [f1, f2]
By concatenating the two features, the combined feature can contain the gender information of both the visible light and thermal images. In Equation (2), the combined feature f is obtained by combing two features (f1 and f2) which are extracted using the visible and thermal image. In the second combination approach, gender classification using visible light and thermal images is performed separately using two SVM classifiers. Using the two SVM classifiers, we obtain the scores for gender classification (decision values of SVM classifiers) with visible and thermal images. These scores are concatenated and input into the third SVM classifier for the score level fusion approach as shown in Figure 3.
The features inputted to the SVM classifiers must be normalized. To do this, we use Z-score normalization method as shown in Equation (3) [16]. In this equation, μ and σ indicate the mean and standard deviation of features in the training database. By using the mean and standard deviation as shown in Equation (3), we normalize the feature and make it similar to the normal distribution:
f = f μ σ
Features extracted by HoG, or MLBP, or feature level fusion have high dimensional characteristic. This characteristic causes difficulty in the classification step based on SVM. In order to solve this problem, our proposed method performs an additional preprocessing step using principal component analysis (PCA). The PCA is a well-known method for dimension reduction and has been used in previous computer vision applications [16]. To do this, we first calculate the covariance matrix C using the extracted image features from the training images using Equation (4):
C = 1 N n = 1 N ( x n μ ) ( x n μ ) T
where, N indicates the total number of training samples, xn is the nth extracted image feature, μ is the mean value of extracted features, and T is the transpose operator. Using the covariance matrix C, we obtain a transformation matrix W by using n eigen-vectors that correspond to the n largest eigen-values of C. The number of eigen-vectors is normally smaller than the dimension of original feature and determined through experiments in which the gender recognition performance is highest. Then, this transformation matrix (W) is used to transform a new input image feature into a low-dimensional feature space [16].
In the final step of our proposed method, we recognize the gender using the SVM method. To do this, we use the OpenCV (ver. 2.4.9) library [26]. SVM classifies input images into male and female classes using support vectors and kernel function. In the general case, the SVM is described as Equation (5). In Equaiton (5), f(x) indicates the classification function; xi is the support vector; and K() is the kernel function that is used to transform the input data into a higher dimensional space. In our experiment, we use two kinds of SVM kernels, linear and RBF, as shown in Equations (6) and (7):
f ( x ) = s i g n ( i = 1 k a i y i K ( x , x i ) + b )
Linear kernel :   K ( x i , x j ) = x i T x j
RBF kernel :   K ( x i , x j ) = e γ x i x j 2
The linear kernel is the simplest kernel and normally used to perform the linear classification problem. And, the RBF kernel is a general kernel that is suitable for most classification problems. Using the training dataset, we can obtain the support vectors as well as the kernel parameters. These support vectors and kernel’s parameters are saved and used for gender classification.

3. Experimental Results

3.1. Description of the Database, Performance Measurement and Experimental Setup

Because there is no public database that contains both visible light and thermal images of the human body, for our experiments we collected our own database using our lab-made device (Figure 4). The database consists of 5852 images of both visible light and thermal images of the human body from 103 persons (66 males and 37 females) with differences in background and body-pose (front, back, side pose). In order to simulate the application of our proposed method in surveillance systems, the dual-camera (in Figure 4) was placed at a height of about 6 m. We captured several images of each person at different times to simulate the effects of body pose. On average, we captured 28 images for each person in different poses. All the images (visible light and thermal images) were captured in an un-controlled environment. Therefore, the illumination and environment greatly affects the captured images. In details, we capture images in both the daytime and nighttime to vary illumination strength. In Figure 7, we show the distribution of image illumination in the visible light and thermal images of our collected database. In previous researches, only the visible images are used for gender recognition [3,12]. Therefore, if the captured image is too dark or too light, detection of the body region could fail and the recognition performance decreases. In this case, the use of thermal images complements the use of visible images. Because the thermal camera captures images using infrared light that is radiated due to the body temperature, the appearance of full body in thermal images is normally distinguished from the background even if the illumination condition is poor (too dark or to bright). As shown in Figure 7, the illumination of visible images in our database contains a large spread from too dark too bright; whereas the illumination of the thermal images is in a smaller brightness range.
Figure 7. The distribution of illumination of images in our collected database: (a) illumination distribution of visible light images; and (b) illumination distribution of thermal images.
Figure 7. The distribution of illumination of images in our collected database: (a) illumination distribution of visible light images; and (b) illumination distribution of thermal images.
Sensors 16 00156 g007
We randomly divided the database into learning and testing sub-databases five times in order to perform a 5-fold cross-validation scheme. In details, the images (both visible light and thermal images) of 53 males and 30 females are used for training, and the remaining images of 13 males and 7 females are used for testing. A detailed description of our collected database is given in Table 2 and Table 3. In order to evaluate the recognition accuracy of our proposed method, we measure the equal error rate (EER) of the training and testing database. By measuring the recognition accuracy of each training-testing time, we obtain the five EER values and the final recognition result of the database is measured as the average value of the five obtained EERs values. The EER is a popular criteria used in biometric systems that indicates the boundary of the recognition system where the false acceptance rate (FAR) is equal to false rejection rate (FRR). For the gender recognition problem, we have two classes of male and female. The FAR value indicates that a female image is falsely recognized as male images; and the FRR value indicates that a male image is falsely recognized as a female image. In Figure 8, we show some examples of visible light and thermal images of the full body that are extracted from our database using the human detection method proposed by Lee et al. [15].
Table 2. Description of visible light and thermal images database of the human body in our experiments.
Table 2. Description of visible light and thermal images database of the human body in our experiments.
DatabaseVisible DatabaseThermal Database(Visible + Thermal) Database
Number of persons103
(66 males/37 females)
103
(66 males/37 females)
103
(66 males/37 females)
Number of Images292629265852
Table 3. Description of learning and testing sub-databases.
Table 3. Description of learning and testing sub-databases.
DatabaseTraining Sub-DatabaseTesting Sub-DatabaseThe Entire Database
Number of persons83
(53 males/30 females)
20
(13 males/7 females)
103
(66 males/37 females)
Figure 8. Examples of visible light and thermal image pairs in our collected database: (a) visible (left)—thermal (right) images of a male in front view; (b) visible (left)—thermal (right) images of a male in back view; (c) visible (left)—thermal (right) images of a female in front views and (d) visible (left)—thermal (right) images of a female in back view.
Figure 8. Examples of visible light and thermal image pairs in our collected database: (a) visible (left)—thermal (right) images of a male in front view; (b) visible (left)—thermal (right) images of a male in back view; (c) visible (left)—thermal (right) images of a female in front views and (d) visible (left)—thermal (right) images of a female in back view.
Sensors 16 00156 g008

3.2. Gender Recognition Using Our Proposed Method

In our first experiment, we measure the recognition accuracy of the recognition system using a single kind of image, either visible light or thermal. When using a single kind of image, the combination step is removed from Figure 1. In this case, the recognition method performs human detection, feature extraction, and recognition using SVM. The use of only visible images corresponds to a conventional system using only visible light images for recognition purpose as in previous studies [3,12]. In addition, by using the thermal images of full human body, our experiments also evaluate the suitability of thermal images for the gender recognition problem. For comparison purposes, we perform experiments using two kinds of feature extraction method as mentioned in Section 2, HoG and MLBP, two kinds of SVM kernels, linear and RBF. The experimental results are shown in Table 4. Also shown in Table 4, the best system performance using only visible images was obtained using the RBF kernel, with EERs of 16.540% (HoG) and 25.088% (MLBP), respectively. Similarly, we obtained the best accuracies (EER) of 19.583% and 20.572% by using the HoG and MLBP feature extraction methods, respectively, using only thermal images. In addition, the recognition accuracy of systems that use the linear kernel of SVM is lower than those that use the RBF kernel. Through these experimental results, we can conclude that the RBF kernel is better suited to gender recognition than the linear kernel. However, the use of the RBF kernel is more complicated than the use of linear kernel, and takes a longer processing time for training and testing than the linear kernel.
As shown in Table 4, the HoG feature extraction method works better than the MLBP feature extraction method by producing lower EER values for both visible and thermal images. This is due to the characteristics of body images and the features extracted by each method. Because the body images have a large variation in poses, hairstyle, clothing and accessories, the gender information is mostly extracted by shape of the body and/or body’s parts. The HoG feature extraction method extracts the strength and the direction of edges on image’s blocks. Therefore, it is efficient for describing the body shape. MLBP on the other hand, extracts the statistical characteristics of the appearance of texture features in an image such as an edge, corner, or spot etc. Due to the variation in body images, the appearance of detailed texture features varies for each individual. Consequently, noise has a greater effect on the MLBP technique than HoG feature extraction.
Table 4. Accuracies of recognition system (EER, FAR vs. GAR) that uses only visible or thermal images of body for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Table 4. Accuracies of recognition system (EER, FAR vs. GAR) that uses only visible or thermal images of body for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
MethodsLinear KernelRBF Kernel
EERFARGAREERFARGAR
Using only visible images for recognitionUsing HoG feature19.63910.00058.09916.54010.00062.758
15.00069.92015.00079.878
19.64280.36416.54283.461
20.00080.44020.00088.051
25.00085.69725.00091.716
Using MLBP feature27.10520.00062.79025.08815.00057.506
25.00069.67620.00066.359
27.10772.89725.09474.918
30.00077.79330.00080.778
35.00081.35935.00086.088
Using only thermal images for recognitionUsing HoG feature23.45915.00060.00219.58310.00057.027
20.00067.19215.00065.791
23.51876.60019.60580.439
25.00079.04920.00080.669
30.00083.49925.00089.815
Using MLBP feature24.00215.00061.90120.57210.00055.929
20.00068.82415.00068.506
24.01976.01420.58079.436
25.00076.77125.00085.236
30.00081.35430.00090.097
Based on the extracted features using visible light and thermal images of human body, in our next experiments we perform the combination of visible and thermal images for gender recognition as shown in Figure 2 and Figure 3. In our second experiment, we combine the two kinds of images using the feature level fusion approach as shown in Figure 2. To do this, the extracted features in each image are concatenated together to form the final feature that contains the gender information in both images. The experimental results are given in Table 5 using the HoG and MLBP feature. Using the HoG feature, the combination of the two kinds of images produce an EER of 15.946%, which is smaller than the EER of 16.540% produced when using only visible images, or EER of 19.583% when using only thermal images. Using the MLBP feature, the recognition accuracy is reduced from 25.088% using visible images, or 20.572% using thermal images, to 18.126% through the feature level fusion approach. These results demonstrate that the combination of two kinds of images can help enhance the accuracy of the recognition system.
Table 5. Recognition accuracy (EER, FAR vs. GAR) of the recognition system using feature level fusion for combining visible and thermal images for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Table 5. Recognition accuracy (EER, FAR vs. GAR) of the recognition system using feature level fusion for combining visible and thermal images for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
MethodLinear KernelRBF Kernel
EERFARGAREERFARGAR
Using HoG feature19.55310.00061.44515.94610.00073.050
15.00070.16915.00080.534
19.60580.49816.19284.299
20.00080.79020.00089.341
25.00084.76425.00092.250
Using MLBP feature21.02215.00063.36618.12610.00060.993
20.00076.48215.00073.669
21.04378.99918.15581.902
25.00083.09620.00084.426
30.00087.07525.00089.898
In the third experiment, we evaluate the recognition performance of the recognition system using the score level fusion approach as depicted in Figure 3. As indicated in Section 2.1, the score level fusion approach requires two layers of SVM classification. The first layer is used to classify the subject gender using the visible light and thermal images separately. This layer outputs scores that indicate the probability of an image belonging to male or female classes. In the second layer, the gender is finally recognized using the scores obtained from the first layer. The detailed results for the HoG and MLBP feature extraction methods are given in Table 6.
As shown in the table, the best accuracy is obtained using the HoG feature extraction method and the RBF kernel in all layers of SVM classification. The EER for the score level combination approach is 14.672%, which is smaller than the errors when using just visible light or thermal images alone (16.540% and 19.583% in Table 4, respectively) or even the feature level fusion approach (15.946% in Table 5). Using the MLBP feature extraction method, the score level fusion produces an EER of 17.642%, which is smaller than the errors produced using single solely visible light or thermal images (25.088% and 20.572% in Table 4, respectively) and the feature level fusion approach (18.126%). Through the experimental results in Table 4, Table 5 and Table 6, we can conclude that the combination of visible and thermal images can help a recognition system enhance its accuracy. In addition, the score level fusion approach outperforms the feature level fusion approach by producing lower recognition error than feature level approach. In Table 7, we summarize the recognition results of our experiments. As shown in this table, the score level fusion method using the HoG feature and RBF kernel (in both layers of SVM classification) produced the best recognition result.
Table 6. Recognition accuracy (EER, FAR vs. GAR) of the recognition system using score level fusion for combining visible and thermal images for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Table 6. Recognition accuracy (EER, FAR vs. GAR) of the recognition system using score level fusion for combining visible and thermal images for gender recognition (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Feature Extraction MethodThe 1st SVM KernelThe 2nd SVM KernelRecognition Accuracy
EERFARGAR
Using HoG featureLinearLinear17.89115.00075.063
17.89282.111
20.00084.353
25.00088.862
RBF17.62815.00075.143
17.66782.411
20.00084.696
25.00089.035
RBFLinear15.15810.00065.884
15.00083.988
15.25484.937
20.00093.900
RBF14.67210.00071.395
14.75485.410
15.00085.410
20.00094.422
Using MLP featureLinearLinear20.83515.00071.802
20.00076.199
21.06879.397
25.00083.019
RBF20.75515.00063.043
20.00076.298
21.06879.559
25.00084.419
RBFLinear17.85710.00059.592
15.00075.281
17.89282.178
20.00085.796
RBF17.64210.00057.442
15.00075.895
17.66782.383
20.00086.253
This result proves that our proposed method is efficient for gender recognition in surveillance systems. In addition, our proposed method outperforms the conventional method that uses single kind of body images (only visible or thermal images) for recognition problem [12].
Table 7. Summary of recognition accuracy (EER, FAR vs. GAR) using a single image type verses the combination of visible and thermal images (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Table 7. Summary of recognition accuracy (EER, FAR vs. GAR) using a single image type verses the combination of visible and thermal images (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Visible Light Image (Conventional Method Using Single Visible Light Images)Thermal Image (Conventional Method Using Single Thermal Images)(Visible Light + Thermal) Images
Feature Level FusionScore Level Fusion (Proposed Method)
EERFARGAREERFARGAREERFARGAREERFARGAR
16.54010.00062.75819.58310.00057.02715.94610.00073.05014.6725.00062.693
15.00079.87815.00065.79115.00080.53410.00071.395
16.54283.46119.60580.43916.19284.29914.75485.410
20.00088.05120.00080.66920.00089.34115.00085.410
25.00091.71625.00089.81525.00092.25020.00094.422
In Figure 9, we show the receiver operating characteristic (ROC) curve for the recognition systems using a single kind of visible/thermal image, feature level fusion, and score level fusion approaches. In this figure, the GAR is the genuine acceptance rate and it is defined as (100–FRR) (%). The figure demonstrates the superior performance of score level fusion methodology when compared to the other investigated methods.
Figure 9. The average ROC curve of recognition systems using different kinds of images.
Figure 9. The average ROC curve of recognition systems using different kinds of images.
Sensors 16 00156 g009
For demonstration purpose, we also show some examples of recognition results using our proposed method against a system that uses only visible and thermal images for the recognition problem in Figure 10 and Figure 11. As shown in Figure 10, although recognition using single images of visible and/or thermal images are fail, the combination of the two kinds of images produce better recognition results. However, the combination still recognizes incorrectly when the input images are too poor in quality, are accompanied by a shadow, or un-normal human body pose as shown in Figure 11.
In the next experiment, we measure the processing time of our proposed method. The results are summarized in Table 8. This experiment was performed using a desktop computer with an Intel Core i7 CPU (3.5 GHz) with 8 GB of RAM. Averagely, it took about 27.5948 ms to process two visible and thermal images and to produce the recognition result. Therefore, we conclude that our proposed method can work in real-time at a speed of up to 36 frames per second (1000/27.5948).
Finally, we use our proposed method to evaluate gender based on only parts of the body. To do this, we divided the entire human body image into three parts: The head, main body, and leg. After that, the recognition accuracies were evaluated using each part of the images. Figure 12 shows the procedure of these experiments.
Figure 10. Examples of recognition result using our proposed method: (a) recognition result of a male; (b,c) recognition results of females.
Figure 10. Examples of recognition result using our proposed method: (a) recognition result of a male; (b,c) recognition results of females.
Sensors 16 00156 g010aSensors 16 00156 g010b
Figure 11. Examples of poor recognition results using our proposed method: (a) images with occlusion; (b) images with un-normal body-pose; and (c) images with poor quality.
Figure 11. Examples of poor recognition results using our proposed method: (a) images with occlusion; (b) images with un-normal body-pose; and (c) images with poor quality.
Sensors 16 00156 g011aSensors 16 00156 g011b
Table 8. Processing time of our proposed method (unit: ms).
Table 8. Processing time of our proposed method (unit: ms).
Human Body DetectionFeature Extraction Using HoG (Two Images)Feature Dimension Reduction (PCA) (Two Images)The 1st SVM LayerThe 2nd SVM LayerTotal
23.13001.63352.75480.07000.006527.5948
The main purpose of these experiments is to exploit the gender information contained in different parts of the human body. Based on previous experimental results, we see that our proposed method produced the best recognition accuracy using the score level fusion method with the RBF kernel. Therefore, in these experiments, we use this fusion approach and the RBF kernel to measure the recognition accuracy of the head, main body and legs. The detail recognition accuracies are shown in Table 9. As shown in Table 9, the head part produces the best recognition accuracy (EER of 15.620%) compared to those of main body part (EER of 20.386%) and leg part (EER of 22.591%) when using the HoG feature extraction method. In addition, the combination of visible and thermal images outperforms the use of solely single visible or thermal images for the purpose of gender recognition. From these results, we can conclude that the parts of head and main body contain more gender information compared to the leg part of human body. The reason is that the leg part contains considerable background noise and limited gender information, such as hair or clothing style when compared to the head or main body parts. Consequently, the recognition accuracy is low and substantially affected by noise compared to other parts of the body.
Figure 12. Procedure of the experiments to evaluate the gender recognition ability of different body parts.
Figure 12. Procedure of the experiments to evaluate the gender recognition ability of different body parts.
Sensors 16 00156 g012
Table 9. The gender recognition performance (EER, FAR vs. GAR) of different parts of human body using our proposed method (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Table 9. The gender recognition performance (EER, FAR vs. GAR) of different parts of human body using our proposed method (the values of GAR and FAR at the EER position are shown in bold-type) (unit: %).
Body’s PartHead PartMain Body PartLeg Part
EERFARGAREERFARGAREERFARGAR
Using Visible Image18.23810.00065.67025.78820.00064.47225.17520.00066.787
15.00075.54825.00073.24225.00074.298
18.24281.76725.79474.21725.25674.906
20.00083.93130.00079.40430.00079.806
25.00088.91335.00085.63035.00084.513
Using Thermal Image22.77915.00066.37626.84520.00062.06627.41420.00061.674
20.00071.56625.00069.66625.00069.402
22.85677.29826.98273.29127.41972.592
25.00082.20030.00077.20330.00075.261
30.00087.84135.00083.99835.00078.912
Our Proposed Method15.62010.00070.66820.38615.00065.92322.59115.00067.983
15.00083.50620.00076.98820.00073.893
15.67984.44020.58079.80822.63177.448
20.00088.99725.00084.18225.00079.688
25.00092.18130.00088.56430.00084.863

4. Conclusions

In this paper, we have proposed a gender recognition method using a combination of visible and thermal images of the human body. Due to the variations in image capturing conditions and the random appearance of the subject due to clothes, accessories, etc. the recognition accuracy of systems that use only single visible images for gender recognition is limited. By using the combination of full body visible and thermal images, we proved that the recognition accuracy is enhanced compared to systems that use only visible or thermal images. In addition, we compared the recognition accuracies when using different feature extraction methods, HoG and MLBP. The experimental results showed that the combination of visible and thermal images produced better recognition accuracy regarding feature extraction methods.
In future work, we plan to enhance the recognition accuracy of this method by counteracting the negative effects such as image’s quality and the effects of background, shadow etc. In addition, we plan to collect a larger database for experiments and evaluate the recognition performance using different recognition methods.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2015R1D1A1A01056761).

Author Contributions

Dat Tien Nguyen and Kang Ryoung Park implemented the overall system of gender recognition with related experiments, and wrote this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ng, C.B.; Tay, Y.H.; Goi, B.M. Recognizing human gender in computer vision: A survey. Lect. Notes Comput. Sci. 2012, 7458, 335–346. [Google Scholar]
  2. Lu, L.; Xu, Z.; Shi, P. Gender Classification of Facial Images Based on Multiple Facial Regions. In Proceedings the World Congress on Computer Science and Information Engineering, Los Angeles, CA, USA, 31 March–2 April 2009; Volume 6, pp. 48–52.
  3. Guo, G.; Mu, G.; Fu, Y. Gender from body: A biologically-inspired approach with manifold learning. Lect. Notes Comput. Sci. 2009, 5996, 236–245. [Google Scholar]
  4. Ichino, M.; Komatsu, N.; Wang, J.-G.; Yun, Y.W. Speaker Gender Recognition Using Score Level Fusion by Adaboost. In Proceedings of the 11th International Conference on Control Automation Robotics and Vision, Singapore, 7–10 December 2010; pp. 648–653.
  5. Sloka, S.; Sridharan, S. Automatic Gender Identification Optimised for Language Independence. In Proceedings of the IEEE Annual Conference on Speech and Image Technologies for Computing and Telecommunications, Brisbane, Australia, 4 December 1997; Volume 1, pp. 145–148.
  6. Arun, K.S.; Rarath, K.S. A Machine Learning Approach for Fingerprint Based Gender Identification. In Proceedings of IEEE Conference on Recent Advances in Intelligent Computational Systems, Trivandrum, India, 22–24 September 2011; pp. 163–167.
  7. Andreu, Y.; Garcia-Sevilla, P.; Mollineda, R.A. Face gender classification: A statistical study when neutral and distorted faces are combined for training and testing purposes. Image Vis. Comput. 2014, 32, 27–36. [Google Scholar] [CrossRef]
  8. Bekios-Calfa, J.; Buenaposada, J.M.; Baumela, L. Robust gender recognition by exploiting facial attributes dependencies. Pattern Recognit. Lett. 2014, 36, 228–234. [Google Scholar] [CrossRef]
  9. Hadid, A.; Pietikainen, M. Combining appearance and motion for face and gender recognition from videos. Pattern Recognit. 2009, 42, 2818–2827. [Google Scholar] [CrossRef]
  10. Li, X.; Zhao, X.; Fu, Y.; Liu, Y. Bimodal Gender Recognition from Face and Fingerprint. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2590–2597.
  11. Xu, X.-Y.; Yu, B.; Wang, Z.; Yin, Z. A Multimodal Gender Recognition Based on Bayesian Hierarchical Model. In Proceedings of the 5th International Conference on Intelligent Human-Machine Systems and Cybernetics, Hangzhou, China, 26–27 August 2013; pp. 410–413.
  12. Cao, L.; Dikmen, M.; Fu, Y.; Huang, T.S. Gender Recognition from Body. In Proceedings of the 16th ACM International Conference on Multimedia, New York, NY, USA, 2008; pp. 725–728.
  13. Tang, J.; Liu, X.; Cheng, H.; Robinette, K.M. Gender Recognition Using 3-D Body Shapes. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2011, 41, 898–908. [Google Scholar] [CrossRef]
  14. Tang, J.; Liu, X.; Cheng, H.; Robinette, K.M. Gender Recognition with Limited Feature Points from 3-D Human Body Shapes. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, Seoul, Korea, 14–17 October 2012; pp. 2481–2484.
  15. Lee, J.H.; Choi, J.-S.; Jeon, E.S.; Kim, Y.G.; Le, T.T.; Shin, K.Y.; Lee, H.C.; Park, K.R. Robust pedestrian detection by combining visible and thermal infrared cameras. Sensors 2015, 15, 10580–10615. [Google Scholar] [CrossRef] [PubMed]
  16. Nguyen, D.T.; Cho, S.R.; Pham, D.T.; Park, K.R. Human age estimation method robust to camera sensor and/or face movement. Sensors 2015, 15, 21898–21930. [Google Scholar] [CrossRef] [PubMed]
  17. C600 Webcam Camera. Available online: https://support.logitech.com/en_us/product/5869 (accessed on 29 December 2015).
  18. Tau2 Thermal Imaging Camera. Available: http://www.flir.com/cores/display/?id=54717 (accessed on 29 December 2015).
  19. Dalal, N.; Triggs, B. Histogram of Oriented Gradients for Human Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 25 June 2005; Volume 1, pp. 886–893.
  20. Nguyen, D.T.; Park, Y.H.; Shin, K.Y.; Kwon, S.Y.; Lee, H.C.; Park, K.R. Fake finger-vein image detection based on fourier and wavelet transforms. Digit. Signal. Prog. 2013, 23, 1401–1413. [Google Scholar] [CrossRef]
  21. Lee, W.O.; Kim, Y.G.; Hong, H.G.; Park, K.R. Face recognition system for set-top box-based intelligent TV. Sensors 2014, 11, 21726–21749. [Google Scholar] [CrossRef] [PubMed]
  22. Nguyen, D.T.; Cho, S.R.; Park, K.R. Human age estimation based on multi-level local binary pattern and regression method. Lect. Notes Electr. Eng. 2014, 309, 433–438. [Google Scholar]
  23. Torrisi, A.; Farinella, G.M.; Puglisi, G.; Battiato, S. Selecting Discriminative CLBP Patterns for Age Estimation. In Proceedings of the IEEE International Conference on Multimedia & Expo Workshops, Turin, Italy, 29 June–3 July 2015; pp. 1–6.
  24. Santarcangelo, V.; Farinella, G.M.; Battiato, S. Gender Recognition: Methods, Dataset and Results. In Proceedings of the IEEE International Conference on Multimedia & Expo Workshops, Turin, Italy, 29 June–3 July 2015; pp. 1–6.
  25. Farinella, G.M.; Farioli, G.; Battiato, S.; Leonardi, S.; Gallo, G. Face re-identification for digital signage applications. Lect. Notes. Compt. Sci. 2014, 8811, 40–52. [Google Scholar]
  26. OpenCV Library. Available online: http://opencv.org/ (accessed on 29 December 2015).

Share and Cite

MDPI and ACS Style

Nguyen, D.T.; Park, K.R. Body-Based Gender Recognition Using Images from Visible and Thermal Cameras. Sensors 2016, 16, 156. https://0-doi-org.brum.beds.ac.uk/10.3390/s16020156

AMA Style

Nguyen DT, Park KR. Body-Based Gender Recognition Using Images from Visible and Thermal Cameras. Sensors. 2016; 16(2):156. https://0-doi-org.brum.beds.ac.uk/10.3390/s16020156

Chicago/Turabian Style

Nguyen, Dat Tien, and Kang Ryoung Park. 2016. "Body-Based Gender Recognition Using Images from Visible and Thermal Cameras" Sensors 16, no. 2: 156. https://0-doi-org.brum.beds.ac.uk/10.3390/s16020156

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop