Skip to main content
Erschienen in: Soft Computing 2/2014

Open Access 01.02.2014 | Methodologies and Application

A study in facial regions saliency: a fuzzy measure approach

verfasst von: Paweł Karczmarek, Witold Pedrycz, Marek Reformat, Elaheh Akhoundi

Erschienen in: Soft Computing | Ausgabe 2/2014

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

People recognize familiar faces in a similar way by using interior facial features (facial regions) such as eyes, nose, mouth, etc. However, the importance of these regions in the realization of face identification and a quantification of the impact of such regions on the recognition process could vary from one region to another. An intuitively appealing observation is that of monotonicity: the more regions are taken into account in the recognition process, the better. From a formal point of view, the relevance of the facial regions and an aggregation of these pieces of experimental evidence can be described in the formal setting of fuzzy measures. Fuzzy measures are of particular interest with this regard given their monotonicity property (which stands in a clear contrast with the more restrictive additivity property inherent to probability–like measures). In this study, we concentrate on the construction of fuzzy measures (more specifically, \( \lambda \)-fuzzy measure) and characterize their performance in the problem of face recognition using a collection of experimental data.
Hinweise
Communicated by G. Acampora.

1 Introduction

Perception and recognition of faces by humans is still a challenging problem. Each individual seems to recognize faces in a slightly different way. Nevertheless, numerous psychological studies report experiments, which confirm that there exist some properties of facial recognition mechanisms (including perception of salient facial regions, their mutual relationships or identification of brain areas responsible for identification of faces) being common for most people. Various methods of automatic face recognition have received significant attention during the recent years mainly because of their broad applications to forensic sciences, border control, passport verification, etc., where computers can help alleviate limitations of humans when working with large and continuously growing collections of data.
Computational face recognition methods can be divided into two main groups, namely holistic matching and feature-based matching methods (cf. Zhao et al. 2003). The first of them concerns a group of methods utilizing information contained in the overall face region such as Eigenfaces (Turk and Pentland 1991), Fisherfaces (Belhumeur et al. 1997), support vector machines (SVMs, Phillips 1998), independent component analysis (ICA, Bartlett et al. 2002), and their modifications, e.g., lattice ICA (Marques and Graña 2012).
The latter group includes methods where the localization and local statistics of facial features like eyes, nose, landmarks, contours etc., are essential. Such approaches include elastic bunch graph matching (EBGM, Wiskott et al. 1997), geometry-oriented methods (Kanade 1977), and local descriptors (Ahonen et al. 2004; Heikkilä et al. 2009).
There are many methods combining these two approaches (so-called hybrid methods), refer to the study reported by Pentland et al. (1994) where the Eigenfaces, Eigen features and the combined modular representation are discussed. Another examples are component-based methods where the face is decomposed into a set of features for which a flexible geometrical relation is allowed to compensate for pose changes (Heisele et al. 2003; Huang et al. 2003; Bonnen et al. 2013). Among these methods, approaches based on fuzzy information fusion produce accurate results, see e.g., Kwak and Pedrycz (2005).
The results concerning a way of perception of faces by humans may be fundamental from the point of view of automatic face recognition. For instance, some facts about significance of face features can play an important role in the design of automatic systems. Generally speaking, faces are processed by humans in a holistic manner (Sinha et al. 2006). Second-order spatial relations (i.e., spacing between features) play a very important role here (Rotshtein et al. 2007). However, internal facial features such as eyes, nose and mouth are relatively more important for recognizing trained (familiar) faces than external facial features (hair and face contour) which are more salient for recognition of untrained (unfamiliar) faces (cf. Ellis et al. 1979; Young et al. 1985). In the work by Davies et al. (1977) the Photofit Kit (a tool developed for reconstruction of faces by the police) was used to change the images of faces. Changes of foreheads, eyes and mouths caused the lowest error rates made by subjects. Similar results were presented in Haig (1986) and Matthews (1978) where the observers indicated that eye/eyebrows followed by mouth and then nose were the most dominant regions as the recognition features (taking into account internal features only). It was shown by O’Donnell and Bruce (2001) that people are highly sensitive to changes in the eye region when they are familiarized with. Other results confirm the significance of the upper half of face (Haig 1986) and eyebrows. In experiment described by Sadr et al. (2003), subjects recognized the faces of celebrities with removed eyebrows significantly worse than the faces without eyes (with the mean difference of 9.5 %). Surveys of works on human recognition of familiar and unfamiliar faces and cue saliency can be found in Johnston and Edmonds (2009) and Shepherd et al. (1981).
Detailed studies on cue saliency in computational face recognition were reported in many studies along with the performance obtained by using numerous techniques. Moreover, it is worth noting that in some situations, especially in crime investigations, we encounter a face image containing only a small visible part of face (for instance, when a person wears a balaclava, a helmet, sunglasses, a veil or a mask, or the head is not aligned properly on the image). Let us describe some results.
Brunelli and Poggio (1993) applied template matching strategy and came up with the following ranking of saliency: eyes, mouth, nose and whole face template. Similar results were obtained in Lam and Yan (1998) where the created feature windows were compared using correlation values as a similarity measure and in Kwak and Pedrycz (2005) where the Fisherfaces method was applied. Radial Basic Functions Networks were used to determine dependency of recognition rate on the facial feature or percent of utilized face image (Sato et al. 1998; Gutta et al. 2002; Gutta and Wechsler 2003). Ekenel and Stiefelhagen (2009) presented a comparison of five salient region-based partitioning approaches and one generic approach with results obtained on five different databases. However, new images were built from the divided segments. Generic partitioning provided the highest correct recognition rates. The best of salient-based partitioning schemes was the one containing five overlapping regions: forehead, left and right eyes, left and right cheeks. Experiments with images consisting of 14 small face regions did not produce such good results. The recognition method was based on feature extraction with discrete cosine transform (Ekenel and Stiefelhagen 2005). Yan and Osadciw (2004) discussed combination of the Eigenfaces method with eyes, mouth, nose and forehead Eigenfeature. Adding an individual Eigenfeature improves the identification accuracy except for the nose Eigenfeatures. In (Dargham et al. 2012) the results of LDA for chosen partial regions were presented. Finally, the performance of 14 facial components in a 3D morphable model approach is considered by Heisele and Blanz (2005). All the results indicate that the eye region exhibit the highest discriminative value. Other works presenting results for particular regions of the face can be found in (Savvides et al. 2004a, b, 2006; Neo et al. 2007, 2010; Teo et al. 2007; Wright et al. 2009; Woodard et al. 2010; Park et al. 2011).
To take advantage of information about the cue saliency in the process of face identification based on an aggregation of a given number of classifiers one can use the fuzzy measure, which may help a determination of weights associated with the corresponding criteria. Then the aggregation procedure can be realized by fuzzy integral. This technique was presented by Kwak and Pedrycz (2005), where the Fisherfaces method was applied to the eye, nose, mouth and whole face regions and by Melin et al. (2005), where the modular neural networks were used to each of the facial areas around eyes, nose and mouths. In the similar way, fuzzy measure and fuzzy integral were applied to aggregate the classifiers obtained by the Fisherface method based on subimages decomposed by wavelets (Kwak and Pedrycz 2004), and to aggregate the separate component SVMs outputs with each component SVM importance (Yan et al. 2006). In the three-dimensional case this method was used by Lee and Marshall (2008). Application of fuzzy measure to gender recognition can be found in (Li et al. 2012). The authors used fuzzy measure to aggregate the results produced by support vector machine classifiers obtained for each of the following features: hair, forehead, eyes, nose, mouth, chin and clothing. Other applications of fuzzy measure in pattern recognition were presented, for instance, in Pedrycz (1990), where the measure was found helpful in the process of features selection. Graves and Nagarajah (2007) presented the model of estimation of the uncertainty for a new observation for multiclass classifier. In Keller et al. (1994) fuzzy measure was applied to fusion of handwritten character classifiers, and in (Yan and Keller 1991) the method of image segmentation was described. All these applications are rooted in the models of decision-making theory. The detailed study of application of the fuzzy measure in it can be found in (Grabisch 1995).
Other methods utilizing the cue saliency and psychophysical mechanisms in the face recognition process were described in (Venkat et al. 2013), where similarity mappings, existing in facial regions by means of Bayesian Networks, were modeled, or in (Da et al. 2010), where LBP-based local descriptor was combined with the weights assigned to distinctive facial areas. Similarly, a 9-region mask was applied to obtain weights for so-called Principal Local Binary Patterns (Pujol and García 2012). Some methods of determining weights for local descriptors based algorithms use human fixations (Fang et al. 2011; Choi et al. 2012).
In this paper, we construct the fuzzy measure by using the results of psychological and computational experiments and relate to the saliency of facial regions for recognizing faces by humans. We are motivated by the fact that the fuzzy measure can potentially capture the important information about the saliency of the particular facial areas and their combinations. People use the information contained in the merged regions and in the particular areas to the recognition purpose. It is intuitively apparent that the more features (with an assumption that each of the features is of relatively high significance itself) are taken into account, the higher performance of face recognition could be achieved. The monotonicity property plays a vital role here. To proceed with a formal description of this effect, we investigate a concept of a fuzzy measure (in which the notion of monotonicity assumes a pivotal role) and provide with extensive experimental evidence in order to quantify the performance of the fuzzy measure. We consider face regions such as eyes, nose, mouth and cheeks areas which are intuitively the most descriptive features of face used in the recognition activities. The main objectives of this study can be outlined as follows:
  • Investigation of the abilities of the fuzzy measure to reflect the importance of information included in facial regions and their aggregates.
  • Quantification of the role of the face regions and, particularly, their combinations when considered in the context of face recognition.
  • Determination of dependencies between potential importance (expressed by fuzzy measure) of merged face areas and recognition rate produced by using well-known face recognition algorithms such as Eigenfaces and Fisherfaces.
  • Comparative analysis of the results obtained in the experiments on automatic and human face recognition.
  • Construction of the Sugeno fuzzy measure using the results of psychological experiments on cue saliency and offering a novel approach to model the mechanism of faces identification by people.
The study provides a comprehensive examination of potential abilities of fuzzy measure to capture and quantify the importance of the information contained in the six of the most salient facial features. We also look at the contributions of all possible concatenations of the facial components when designing face classifiers. This way we concentrate on the face identification when taking into account both the information contained in the entire face (as a result of features merging) and in the local components of the face. We examine the way on how the presence of these facial components influences the recognition process.
The paper is organized as follows. The general processing scheme is presented in Sect. 2. In Sect. 3, we discuss fuzzy measure and its usage and elaborate on the semantics in the context of feature-based face recognition. In Sect. 4, presented are experiments while conclusions are covered in Sect. 5.

2 A general processing scheme

An overall scheme highlighting a sequence of main processing phases is presented in Fig. 1. First, a face image is preprocessed (which includes cropping, scaling, and eventual histogram equalization). In the sequel, a position of salient facial regions such as eyes, eyebrows, nose, mouth and cheeks are determined manually by selection of facial areas giving the highest accuracy rates when running some preliminary tests. Then, using the selected facial segments, we determine the accuracies associated with all the regions, i.e., the combinations of these atomic areas by applying the PCA method and, simultaneously, PCA followed by the well-known LDA dimensionality reduction procedure (known as Fisherfaces (Belhumeur et al. 1997)) and determining the recognition rates. The Euclidean distance function is commonly used.
Using the accuracy values obtained for the atomic facial regions we construct a fuzzy measure (more specifically, \( \lambda \)-fuzzy measure) for all the combinations of these regions. In parallel, we determine the recognition rates obtained for the some combinations of the areas when using the PCA and Fisherfaces methods.

3 Interpretation of the fuzzy measure

Classifying face images based on a given number of classifiers (i.e., face regions compared independently), we take into account weights of criteria (both individually and considering their groups, i.e. concatenations of facial regions). They express various qualities of recognition obtained from the classifiers and should be determined to affect the final decision about the classification of an unknown image in a proper way. These weights being used in the aggregation process of outcomes of different classifiers can be represented by fuzzy measure and the values obtained in the process described in previous section, see Fig. 1, are suitable for this purpose. Moreover, fuzzy measure can express the dependencies between these regions on a basis of which classifiers are constructed (Grabisch 1995). More formally, let us assume that \( X = \left\{ {x_{1} , \ldots ,x_{n} } \right\} \) denote the overall face area, where \( x_{1} , \ldots ,x_{n} \) stand for non-overlapping facial segments such as eyes, nose, etc. A fuzzy measure is defined as a set function \( g:P\left( X \right) \to \left[ {0,1} \right] \) satisfying the following conditions:
1.
\( g\left( \emptyset \right) = 0,g\left( X \right) = 1; \)
 
2.
\( g\left( A \right) \le g\left( B \right) \) for \( A \subset B,\,\,\,\text{where}\,\,A, B \in P\left( X \right). \)
 
The first condition quantifies with the observation that having the entire face image we have complete information about the face. The second property (monotonicity) quantifies the psychologically motivated observation that the likelihood of a proper identification of the individual increases when the knowledge about the available region of face is augmented by pieces of knowledge concerning other facial areas. In the original definition of the fuzzy measure the limit condition is also provided (Sugeno 1974), \( \mathop {\lim }\nolimits_{n \to \infty } g\left( {A_{n} } \right) = g\left( {\mathop {\lim }\nolimits_{n \to \infty } A_{n} } \right), \) where \( \left\{ {A_{n} } \right\},\;n = 1, \, 2, \ldots , \) is an arbitrary increasing sequence of measurable sets.
Sugeno (1974) proposed a parametric version of the fuzzy measure (often denoted \( g_{\lambda } ) \) by introducing the following aggregation scheme.
$$ g\left( {A\mathop \cup \nolimits B} \right) = g\left( A \right) + g\left( B \right) + \lambda g\left( A \right)g\left( B \right),\quad \lambda > - 1, $$
(1)
to be satisfied for any pair of disjoint sets \( A \) and \( B \). The value of the parameter \( \lambda \) describes the dependency between the two combined face regions. Note that if \( \lambda\, < \,0 \) then the measure has the property of sub-additivity which means that the satisfaction arising from one source of evidence, i.e. region, more or less entails the satisfaction arising from the second and they are in competition (or redundancy). Here the combination of the areas can be not as efficient as one might have expected. On the other hand, the values \( \lambda > 0 \) imply a synergy effect meaning that these sources of evidence support each other (Grabisch 1995; Pedrycz and Gomide 1998). In such a case a combination of two or more classifiers should be more efficient. If the set of facial regions is chosen, it is obvious that the Sugeno measure cannot be both sub-additive for one group of parts of face and super-additive for another because of the constant value of the parameter \( \lambda \).
The parameter \( \lambda \) can be determined by solving a polynomial equation of the following form (Sugeno 1974):
$$ 1 + \lambda = \mathop \prod \limits_{i = 1}^{n} \left( {1 + \lambda g_{i} } \right), \,g_{i} = g\left( {\left\{ {x_{i} } \right\}} \right), $$
where, as previously, \( x_{1} , \ldots ,x_{n} \) are non-overlapping facial regions such as eyes, nose, etc. There exists a unique solution \( \lambda \) to the above equation with \( \lambda > - 1, \;\lambda \ne 0 \) (Sugeno 1974). The values \( g_{i} \) are known and are called densities of the fuzzy measure. Denoting \( A_{i} = \left\{ {x_{1} , \ldots ,x_{i} } \right\},A_{i + 1} = \left\{ {x_{1} , \ldots ,x_{i} ,x_{i + 1} } \right\}, \) we use the recurrence formula to calculate the fuzzy measure over the combined facial regions:
$$ g\left( {A_{i + 1} } \right) = g\left( {A_{i} } \right) + g_{i + 1} + \lambda g\left( {A_{i} } \right)g_{i + 1} $$
(2)
with \( g\left( {A_{1} } \right) = g_{1} \).

4 Experiments

The main objective of the series of experiments carried out in this study is to determine the accuracies of recognition for salient facial regions and their combinations as well as the corresponding values of fuzzy measure for these areas. Similarly, we calculate these values using the results of psychological studies reported in the literature. We are interested in the examination of the properties of fuzzy measures obtained this way. The results of experiments reported in this study are obtained for the AT&T (formerly ORL) image database (http://​www.​cl.​cam.​ac.​uk/​research/​dtg/​attarchive/​facedatabase.​html) and the Facial Recognition Technology (FERET) Database (Phillips et al. 1998).
The AT&T database consists of 400 images of 40 individuals with various illumination, pose, and expression. The number of images per subject is always 10. In the experiments completed for the FERET database we use its sets (called ba, bk and bj) which consist of 600 images of 200 individuals (3 pictures of each subject) taken for various expression and different illumination conditions.
In the preliminary tests, the following six facial areas were selected: eyebrows, eyes, nose, mouth, left cheek, and right cheek. These regions cover most of the face area and are important in the process of recognition/classification realized by a human being. Next, we found the detailed sizes of the regions, which produce the highest accuracy rates for each of these six segments (refer to Fig. 2; Table 1 for details).
Table 1
Face regions and their characteristics (size of the regions)
Region
AT&T database
FERET database
Width
Height
Width
Height
Original face image
92
112
256
384
Cropped face image
90
94
100
140
Eyebrows
88
14
91
14
Eyes
82
14
84
15
Nose
35
28
37
31
Mouth
51
28
54
29
Cheeks (left and right)
22
55
24
72
In the first series of experiments we present the results of recognition rates and calculated fuzzy measures for the atomic regions and their combinations by using the PCA method. We divide the set of images from the AT&T database taking five randomly chosen images of an individual to the training set and the rest of images are included to the testing set. Similar experiments we do with images from the FERET database (two training images and one testing image per person). Simultaneously, we apply the Fisherfaces method to the same datasets. Each of these computations was repeated 100 times and the final recognition rate is taken as an average of all the results.
The values of the recognition rates for all the atomic salient facial regions are reported in Table 2. Figure 3a illustrates the values of resulting from the calculations of the fuzzy measure (1) and (2) with respect to combinations of regions and the corresponding recognition accuracy values obtained for the concatenated images (i.e., vectors being the result of merging two images treated as vectors of pixel values) of two facial segments obtained by using the PCA and PCA followed by LDA methods on the AT&T database, PCA and PCA followed by LDA on the FERET database. Figure 3b–d present similar standings for combinations of three, four, five and all the regions, respectively.
Table 2
Recognition rates (%) obtained for salient atomic facial regions
Region
AT&T database
FERET database
PCA
PCA + LDA
PCA
PCA + LDA
Eyebrows
62.16
81.75
28.81
72.93
Eyes
67.03
79.86
15.42
52.4
Nose
59.77
66.23
10.29
31.28
Mouth
49.31
60.89
4.08
18.75
Left cheek
36.55
68.06
9.4
30.68
Right cheek
39.42
67.96
10
36.17
The scatter plot of the values of accuracy and the accuracy values produced when using the fuzzy measure are presented in Fig. 4a–d. Along with the results, we also show a linear regression, which expresses the relationship between the results formed by the fuzzy measure and those obtained when running classification schemes. Note that the accuracy values have been rescaled to be consistent with the boundary condition imposed on the fuzzy measure.
Similar results are obtained for the results produced when using of probability measure, representing the class of measures fulfilling the condition of additivity (see Fig. 4e–h). This measure is constructed using the recognition rates obtained for basic facial segments (eyes, nose, etc.) and then simply applying the additivity condition for the combination of regions. All the values are normalized to fulfill the boundary condition 1.
The values of the parameter \( \lambda \), maximal and minimal differences between the values of the Sugeno measure and the recognition rate obtained in the classification process are presented in Table 3. Figure 5 visualizes the recognition rates of the Fisherfaces method and corresponding fuzzy measures for the two groups consisting combinations of four facial parts, i.e., eyebrows, eyes, nose, and mouth and nose, mouth, left cheek, and right cheek. These parts are distinguished because they correspond to the upper and lower parts of face in the first and in the second case, respectively. In Fig. 6 we include the values of the classification accuracy along with the values of the fuzzy measure for the areas of eyes and mouths being gradually augmented by other regions proceeding with their neighbors.
Table 3
The values of the parameter of the gλ fuzzy measure, maximal and minimal differences between the calculated fuzzy measure and recognition rates obtained when using merged facial regions
Method
Parameter \( \lambda \)
Minimal difference
Maximal difference
PCA (AT&T)
−0.98944
0.11
0.31
PCA + LDA (AT&T)
−0.9995
0.06
0.17
PCA (FERET)
0.82529
0.04
0.78
PCA + LDA (FERET)
−0.9608
0.02
0.24
The results confirm in general the following facts: For any part of the face serving as a classifier (especially containing eyes) the recognition accuracy is better when the area of consideration is getting bigger. Moreover, the recognition rate increases when the combination of salient facial regions is considered. Finally, the most descriptive region of the face is eyes and eyebrows area (in general the upper half of face). Especially eyebrows are very significant in the process of computational face identification (over 81 % accuracy rate obtained for the AT&T database) and their presence in the considered area can increase the recognition rate significantly (see Fig. 6b). Even augmenting the eyes region by the other face segments does not increase the recognition rate in a meaningful fashion (see Fig. 6a).
What becomes even more important, is that the results show that the fuzzy measure is strongly sub-additive (\( \lambda \le - 0.9608 \)); this occurs in all considered cases with an exception of PCA for FERET. In this case the accuracies are very low and the method is rather inefficient. Therefore, the value of the parameter λ is positive (it comes from the boundary conditions on fuzzy measure). However, in all the considered cases fuzzy measure can be treated as a good source of evidence for the salient facial regions as it corresponds to the accuracies obtained for the combinations of regions. This property leads to the conclusion that the interactions between crucial segments may be reflected by Sugeno fuzzy measure. It is easily seen from the figures presenting the scatter between real recognition rates and values of Sugeno fuzzy measure. The points lying far from the regression line are the points corresponding to atomic facial regions or the combinations of segments underrating the recognition accuracies, e.g. cheeks or nose areas. Moreover, the correlation between fuzzy measure and accuracy when all the six crucial facial segments (eyebrows, eyes without eyebrows, nose, mouth, left and right cheeks) are considered is less than the correlation in the cases of four facial regions (see Fig. 5). Therefore, we conclude that in the situation when the chosen segments of face are occluded it may be sufficient to take into account only the most important facial parts such as eyes and eyebrows with combinations with others being available for recognition. Nevertheless, the general trend is that both real recognition accuracy and Sugeno fuzzy measure increase when the higher number of facial regions is merged. The highest differences between the classification accuracy and fuzzy measure can be observed for the regions placed at the lower part of face such as nose, mouth and cheeks. As it was discussed in Sect. 1, these regions are considered to be less useful as classifiers than the eyes area. It can be noticed that fuzzy measure slightly tends to overvalue the potential weight of information included in these regions.
Comparing the scatter between the accuracy of classifiers and fuzzy and probability measures, respectively, it is easy to see that fuzzy measure is more flexible and its values better fit to the scaled values of recognition rates than the values of the measure fulfilling the condition of additivity. Only in the case of PCA on the FERET database (where the method is of low efficiency) the scatters are similar (see Fig. 4c, g).
Now let us consider the psychological experiments reported by Matthews (1978). In these experiments, the subjects were to answer a question about similarity or dissimilarity of two images after modifications of one or more facial features: eyes, eyebrows, nose, mouth, chin and hair. The last feature is not discussed here as an exterior face area. The face images were constructed using a police “Identikit” from the transparent overlays of facial features. Chosen results (and Sugeno measure values) are presented in Fig. 7. A comparison of accuracies observed in this experiment and in our computational experiments with the Fisherfaces method for chosen facial regions is presented in Fig. 8a. Similar comparison containing the values of the Sugeno fuzzy measure is presented in Fig. 8b.
The value of parameter \( \lambda \) obtained from the accuracies of human recognition is −0.99994 and the correlation between recognition accuracies and computed Sugeno measures for selected parts of face is 0.846. It means, as in the case of automatic face identification described above, that this measure reflects the way people recognize faces and can be applied to model the interactions between facial segments. Figure 8a shows that in the process of identifying faces by humans, particular features and their combinations have similar meaning and their saliency is comparable with the saliency in the process of automatic face recognition (keeping the relationships between their values). As a consequence similar situation takes place in case of the fuzzy measure values (see Fig. 8b).
In the last series of experiments we divide the group of individuals into eight subsets \( A_{1} , \ldots , \;A_{8} ,\; A_{1} \subset A_{2} \subset \cdots \subset A_{8} . \) In case of the AT&T database first of this sets consists of 5 individuals, second consists of 10, etc. Similarly, the subsets of FERET are built from images of 25, 50,…, 200 people, respectively. As before, we find recognition accuracies using the PCA and PCA + LDA methods. Next, we construct the Sugeno fuzzy measure taking as fuzzy densities the accuracies for the atomic facial segments such as eyebrows, eyes, nose, mouth, left and right cheeks areas. The values of the parameter \( \lambda \) are presented in Fig. 9. It can be observed that this value tends to −1 while the number of people in the considered dataset decreases and while the method is more efficient, i.e. having highest accuracies. The reason of this is the boundary condition 1. The measure tends to fulfill it by overvaluation of the results. The most meaningful example here is almost linear dependency between the number of people in the dataset and the values of \( \lambda \) in case of the PCA method.
Figure 10 presents the values of fuzzy measure depending on number of considered classes from each database. Four upper and four lower combined facial segments were taken under consideration. The results show that the measure is rather stable in the case of efficient method such as Fisherfaces with five training images per class for AT&T database or in the case of significant facial area, e.g. eyes and their neighborhood. However, in other cases, particularly PCA, the value of fuzzy measure decreases when the number of classes increases. It is strictly related with the real recognition rates whose values decrease in a similar way.

5 Conclusions and future work

In this study, we have investigated an application of the fuzzy measure (Sugeno fuzzy measure) as a vehicle to quantify a way of aggregation of important discriminatory information conveyed by facial regions. We discussed the properties of additivity and monotonicity in the context of face recognition based on the salient facial regions. The comprehensive series of experiments led us to the conclusion that the fuzzy measure can be sought as a sound vehicle to aggregate evidence—pieces of knowledge residing within face segments. In most cases, we can conclude that the fuzzy measure (owing to its monotonicity) comes as sound classification model.
Future work may include an efficient application of the fuzzy measure (particularly, related to the psychological studies) in face recognition systems based on other than PCA or LDA methods, development of measure being more flexible in the sense of expressing the interactions between higher number of the facial features, as well as insightful study of the way of determining membership grades of a class in a classifier playing a significant role in information fusion by fuzzy integral. Other interesting issue may be a deepened study of the eyes and eyebrows area region and its impact on the recognition process. Finding the subareas highly responsible for the quality of the recognition process would significantly reduce the dimensionality of data needed to the computation.

Acknowledgments

This work was supported by the Natural Sciences and Engineering Research Council (NSERC) through a Strategic Research Grant.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
Literatur
Zurück zum Zitat Ahonen T, Hadid A, Pietikäinen M (2004) Face recognition with local binary patterns. In: Proceedings of the 8th European conference on computer vision. LNCS 3021:469–481. doi:10.1007/978-3-540-24670-1_36 Ahonen T, Hadid A, Pietikäinen M (2004) Face recognition with local binary patterns. In: Proceedings of the 8th European conference on computer vision. LNCS 3021:469–481. doi:10.​1007/​978-3-540-24670-1_​36
Zurück zum Zitat Belhumeur PN, Hespanha JP, Kriegman DJ (1997) Eigenfaces vs. Fisherfaces: recognition using class specific linear projection. IEEE Trans Pattern Anal Mach Intell 19:711–720. doi:10.1109/34.598228 CrossRef Belhumeur PN, Hespanha JP, Kriegman DJ (1997) Eigenfaces vs. Fisherfaces: recognition using class specific linear projection. IEEE Trans Pattern Anal Mach Intell 19:711–720. doi:10.​1109/​34.​598228 CrossRef
Zurück zum Zitat Choi E, Lee S-W, Wallraven C (2012) Face recognition with enhanced local Gabor binary pattern from human fixations. In: Proceedings of 2012 IEEE International conference on systems, man, and cybernetics (SMC), pp 863–867. doi:10.1109/ICSMC.2012.6377836 Choi E, Lee S-W, Wallraven C (2012) Face recognition with enhanced local Gabor binary pattern from human fixations. In: Proceedings of 2012 IEEE International conference on systems, man, and cybernetics (SMC), pp 863–867. doi:10.​1109/​ICSMC.​2012.​6377836
Zurück zum Zitat Da B, Sang N, Li C (2010) Face recognition by estimating facial distinctive information distribution. In: Zha H, Taniguchi RI, Maybank S (eds) Computer Vision—ACCV 2009, Part III, LNCS 5996: 570–580. doi:10.1007/978-3-642-12297-2_55 Da B, Sang N, Li C (2010) Face recognition by estimating facial distinctive information distribution. In: Zha H, Taniguchi RI, Maybank S (eds) Computer Vision—ACCV 2009, Part III, LNCS 5996: 570–580. doi:10.​1007/​978-3-642-12297-2_​55
Zurück zum Zitat Dargham JA, Chekima A, Hamdan M (2012) Hybrid component-based face recognition system. In: Omatu S et al. (eds) Distributed computing and artificial intelligence, Advances in Intelligent and Soft Computing 151: 573–580. doi:10.1007/978-3-642-28765-7_69 Dargham JA, Chekima A, Hamdan M (2012) Hybrid component-based face recognition system. In: Omatu S et al. (eds) Distributed computing and artificial intelligence, Advances in Intelligent and Soft Computing 151: 573–580. doi:10.​1007/​978-3-642-28765-7_​69
Zurück zum Zitat Ekenel HK, Stiefelhagen R (2005) Local appearance based face recognition using discrete cosine transform. In: Proceedings of the 13th European Signal Processing Conference (EUSIPCO). Antalya, pp 2484–2488 Ekenel HK, Stiefelhagen R (2005) Local appearance based face recognition using discrete cosine transform. In: Proceedings of the 13th European Signal Processing Conference (EUSIPCO). Antalya, pp 2484–2488
Zurück zum Zitat Ekenel HK, Stiefelhagen R (2009) Generic versus salient region-based partitioning for local appearance face recognition. In: Tistarelli M, Nixon MS (eds) Advances in biometrics, LNCS 5558: 367–375. doi:10.1007/978-3-642-01793-3_38 Ekenel HK, Stiefelhagen R (2009) Generic versus salient region-based partitioning for local appearance face recognition. In: Tistarelli M, Nixon MS (eds) Advances in biometrics, LNCS 5558: 367–375. doi:10.​1007/​978-3-642-01793-3_​38
Zurück zum Zitat Ellis HD, Shepherd JW, Davies GM (1979) Identification of familiar and unfamiliar faces from internal and external features: some implications for theories of face recognition. Perception 8:431–439. doi:10.1068/p080431 CrossRef Ellis HD, Shepherd JW, Davies GM (1979) Identification of familiar and unfamiliar faces from internal and external features: some implications for theories of face recognition. Perception 8:431–439. doi:10.​1068/​p080431 CrossRef
Zurück zum Zitat Fang F, Qing L, Wang C, Miao J, Chen X, Gao W (2011) Attention driven face recognition, learning from Human Vision System. Int J Comput Sci Issues 8:8–13 Fang F, Qing L, Wang C, Miao J, Chen X, Gao W (2011) Attention driven face recognition, learning from Human Vision System. Int J Comput Sci Issues 8:8–13
Zurück zum Zitat Gutta S, Wechsler H (2003) Partial faces for face recognition: Left vs right half. In: Petkov N, Westenberg MA (eds) Computer analysis of images and patterns, LNCS 2756: 630–637. doi:10.1007/978-3-540-45179-2_77 Gutta S, Wechsler H (2003) Partial faces for face recognition: Left vs right half. In: Petkov N, Westenberg MA (eds) Computer analysis of images and patterns, LNCS 2756: 630–637. doi:10.​1007/​978-3-540-45179-2_​77
Zurück zum Zitat Gutta S, Philomin V, Trajković M (2002) An investigation into the use of partial faces for face recognition. In: Proceedings of the 5th IEEE International conference on automatic face and gesture recognition, 2002, pp 28–33. doi:10.1109/AFGR.2002.1004126 Gutta S, Philomin V, Trajković M (2002) An investigation into the use of partial faces for face recognition. In: Proceedings of the 5th IEEE International conference on automatic face and gesture recognition, 2002, pp 28–33. doi:10.​1109/​AFGR.​2002.​1004126
Zurück zum Zitat Heisele B, Blanz V (2005) Morphable models for training a component-based face recognition system. In: Zhao W, Chellapa R (eds) Face processing, advanced modeling and methods. Elsevier, London, pp 439–462 Heisele B, Blanz V (2005) Morphable models for training a component-based face recognition system. In: Zhao W, Chellapa R (eds) Face processing, advanced modeling and methods. Elsevier, London, pp 439–462
Zurück zum Zitat Huang J, Heisele B, Blanz V (2003) Component-based face recognition with 3D morphable models. In: Audio- and video-based biometric person authentication, LNCS 2688: 27–34. doi:10.1007/3-540-44887-X_4 Huang J, Heisele B, Blanz V (2003) Component-based face recognition with 3D morphable models. In: Audio- and video-based biometric person authentication, LNCS 2688: 27–34. doi:10.​1007/​3-540-44887-X_​4
Zurück zum Zitat Kanade T (1977) Computer recognition of human faces. Birkhauser, BaselCrossRef Kanade T (1977) Computer recognition of human faces. Birkhauser, BaselCrossRef
Zurück zum Zitat Lam K-M, Yan H (1998) An analytic-to-holistic approach for face recognition based on a single frontal view. IEEE Trans Pattern Anal Mach Intell 20:673–686. doi:10.1109/34.689299 CrossRef Lam K-M, Yan H (1998) An analytic-to-holistic approach for face recognition based on a single frontal view. IEEE Trans Pattern Anal Mach Intell 20:673–686. doi:10.​1109/​34.​689299 CrossRef
Zurück zum Zitat Melin P, Felix C, Castillo O (2005) Face recognition using modular neural networks and the fuzzy Sugeno integral for response integration. Int J Intell Syst 20:275–291. doi:10.1002/int.v20:2 CrossRef Melin P, Felix C, Castillo O (2005) Face recognition using modular neural networks and the fuzzy Sugeno integral for response integration. Int J Intell Syst 20:275–291. doi:10.​1002/​int.​v20:​2 CrossRef
Zurück zum Zitat Neo HF, Teo CC, Teoh ABJ (2007) A study on optimal face ratio for recognition using part-based feature extractor. In: Proceedings of 3rd International IEEE conference on signal-image technologies and internet-based system, 2007. SITIS ‘07, pp 735–741. doi:10.1109/SITIS.2007.52 Neo HF, Teo CC, Teoh ABJ (2007) A study on optimal face ratio for recognition using part-based feature extractor. In: Proceedings of 3rd International IEEE conference on signal-image technologies and internet-based system, 2007. SITIS ‘07, pp 735–741. doi:10.​1109/​SITIS.​2007.​52
Zurück zum Zitat Neo HF, Teo CC, Teoh ABJ (2010) Development of partial face recognition framework. In: Proceedings of 7th International conference on computer graphics, imaging and visualization (CGIV), 2010, pp 142–146. doi:10.1109/CGIV.2010.29 Neo HF, Teo CC, Teoh ABJ (2010) Development of partial face recognition framework. In: Proceedings of 7th International conference on computer graphics, imaging and visualization (CGIV), 2010, pp 142–146. doi:10.​1109/​CGIV.​2010.​29
Zurück zum Zitat O’Donnell C, Bruce V (2001) Familiarisation with faces selectively enhances sensitivity to changes made to the eyes. Perception 30:755–764. doi:10.1068/p3027 CrossRef O’Donnell C, Bruce V (2001) Familiarisation with faces selectively enhances sensitivity to changes made to the eyes. Perception 30:755–764. doi:10.​1068/​p3027 CrossRef
Zurück zum Zitat Pedrycz W, Gomide F (1998) An introduction to fuzzy sets: analysis and design. The MIT Press, CambridgeMATH Pedrycz W, Gomide F (1998) An introduction to fuzzy sets: analysis and design. The MIT Press, CambridgeMATH
Zurück zum Zitat Pentland A, Moghaddam B, Starner T (1994) View-based and modular Eigenspaces for face recognition. In: Proceedings CVPR ‘94, 1994 IEEE Computer Society conference on computer vision and pattern recognition, pp 84–91. doi:10.1109/CVPR.1994.323814 Pentland A, Moghaddam B, Starner T (1994) View-based and modular Eigenspaces for face recognition. In: Proceedings CVPR ‘94, 1994 IEEE Computer Society conference on computer vision and pattern recognition, pp 84–91. doi:10.​1109/​CVPR.​1994.​323814
Zurück zum Zitat Phillips PJ (1998) Support vector machines applied to face recognition. Adv Neural Inf Process Syst 11:803–809 Phillips PJ (1998) Support vector machines applied to face recognition. Adv Neural Inf Process Syst 11:803–809
Zurück zum Zitat Rotshtein P, Geng JJ, Driver J, Dolan RJ (2007) Role of features and second-order spatial relations in face discrimination, face recognition, and individual face skills: behavioral and functional magnetic resonance imaging data. J Cogn Neurosci 19:1435–1452. doi:10.1162/jocn.2007.19.9.1435 CrossRef Rotshtein P, Geng JJ, Driver J, Dolan RJ (2007) Role of features and second-order spatial relations in face discrimination, face recognition, and individual face skills: behavioral and functional magnetic resonance imaging data. J Cogn Neurosci 19:1435–1452. doi:10.​1162/​jocn.​2007.​19.​9.​1435 CrossRef
Zurück zum Zitat Sato K, Shah S, Aggarwal JK (1998) Partial face recognition using radial basis function networks. In: Proceedings of 3rd IEEE International conference on automatic face and gesture recognition, pp 288–293. doi:10.1109/AFGR.1998.670963 Sato K, Shah S, Aggarwal JK (1998) Partial face recognition using radial basis function networks. In: Proceedings of 3rd IEEE International conference on automatic face and gesture recognition, pp 288–293. doi:10.​1109/​AFGR.​1998.​670963
Zurück zum Zitat Savvides M, Kumar BVKV, Khosla PK (2004) “Corefaces”- robust shift invariant PCA based correlation filter for illumination tolerant face recognition. In: Proceedings of the 2004 IEEE Computer Society conference on computer vision and pattern recognition, CVPR, pp 834–841. doi:10.1109/CVPR.2004.1315251 Savvides M, Kumar BVKV, Khosla PK (2004) “Corefaces”- robust shift invariant PCA based correlation filter for illumination tolerant face recognition. In: Proceedings of the 2004 IEEE Computer Society conference on computer vision and pattern recognition, CVPR, pp 834–841. doi:10.​1109/​CVPR.​2004.​1315251
Zurück zum Zitat Savvides M, Kumar BVKV, Khosla PK (2004) Eigenphases vs. Eigenfaces. In: Proceedings of the 17th International conference on pattern recognition, ICPR, pp 810–813. doi:10.1109/ICPR.2004.1334652 Savvides M, Kumar BVKV, Khosla PK (2004) Eigenphases vs. Eigenfaces. In: Proceedings of the 17th International conference on pattern recognition, ICPR, pp 810–813. doi:10.​1109/​ICPR.​2004.​1334652
Zurück zum Zitat Savvides M, Abiantun R, Heo J, Park S, Xie C, Vijayakumar BVK (2006) Partial and holistic face recognition on FRGC-II data using support vector machine kernel correlation feature analysis. In: Proceedings of CVPRW ‘06 conference on computer vision and pattern recognition workshop. doi:10.1109/CVPRW.2006.153 Savvides M, Abiantun R, Heo J, Park S, Xie C, Vijayakumar BVK (2006) Partial and holistic face recognition on FRGC-II data using support vector machine kernel correlation feature analysis. In: Proceedings of CVPRW ‘06 conference on computer vision and pattern recognition workshop. doi:10.​1109/​CVPRW.​2006.​153
Zurück zum Zitat Shepherd J, Davies G, Ellis H (1981) Studies of cue saliency. In: Davies G, Ellis HD, Shepherd JW (eds) Perceiving and remembering faces. Academic Press, New York, pp 105–131 Shepherd J, Davies G, Ellis H (1981) Studies of cue saliency. In: Davies G, Ellis HD, Shepherd JW (eds) Perceiving and remembering faces. Academic Press, New York, pp 105–131
Zurück zum Zitat Sugeno M (1974) Theory of fuzzy integral and its applications. Dissertation, Tokyo Institute of Technology Sugeno M (1974) Theory of fuzzy integral and its applications. Dissertation, Tokyo Institute of Technology
Zurück zum Zitat Teo CC, Neo HF, Teoh ABJ (2007) A study on partial face recognition of eye region. In: Proceedings of International conference on machine vision, 2007, ICMV 2007, pp 46–49. doi:10.1109/ICMV.2007.4469271 Teo CC, Neo HF, Teoh ABJ (2007) A study on partial face recognition of eye region. In: Proceedings of International conference on machine vision, 2007, ICMV 2007, pp 46–49. doi:10.​1109/​ICMV.​2007.​4469271
Zurück zum Zitat Woodard DL, Pundlik SJ, Lyle JR, Miller PE (2010) Periocular region appearance cues for biometric identification. In: Proceedings of IEEE Computer Society conference on computer vision and pattern recognition workshops (CVPRW), 2010, pp 162–169. doi:10.1109/CVPRW.2010.5544621 Woodard DL, Pundlik SJ, Lyle JR, Miller PE (2010) Periocular region appearance cues for biometric identification. In: Proceedings of IEEE Computer Society conference on computer vision and pattern recognition workshops (CVPRW), 2010, pp 162–169. doi:10.​1109/​CVPRW.​2010.​5544621
Zurück zum Zitat Yan B, Keller J (1991) Conditional fuzzy measures and image segmentation. In: Proceedings of NAFIPS-91. NAFIPS 1981–1991: a decade of growth in uncertainty modeling, University of Missouri-Columbia, Missouri, pp 32–36 Yan B, Keller J (1991) Conditional fuzzy measures and image segmentation. In: Proceedings of NAFIPS-91. NAFIPS 1981–1991: a decade of growth in uncertainty modeling, University of Missouri-Columbia, Missouri, pp 32–36
Zurück zum Zitat Yan Y, Osadciw LA (2004) Intra-difference based segmentation and face identification. In: Jain AK, Ratha NK (eds) Biometric technology for human identification. Proceedings of SPIE 5404, pp 502–510 Yan Y, Osadciw LA (2004) Intra-difference based segmentation and face identification. In: Jain AK, Ratha NK (eds) Biometric technology for human identification. Proceedings of SPIE 5404, pp 502–510
Zurück zum Zitat Yan G, Ma G, Zhu L (2006) Support vector machines ensemble based on fuzzy integral for classification. In: Advances in neural networks—ISNN 2006, LNCS 3971:974–980. doi:10.1007/11759966_143 Yan G, Ma G, Zhu L (2006) Support vector machines ensemble based on fuzzy integral for classification. In: Advances in neural networks—ISNN 2006, LNCS 3971:974–980. doi:10.​1007/​11759966_​143
Zurück zum Zitat Young AW, Hay DC, McWeeny KH, Flude BM, Ellis AW (1985) Matching familiar and unfamiliar faces on internal and external features. Perception 14:737–746. doi:10.1068/p140737 CrossRef Young AW, Hay DC, McWeeny KH, Flude BM, Ellis AW (1985) Matching familiar and unfamiliar faces on internal and external features. Perception 14:737–746. doi:10.​1068/​p140737 CrossRef
Metadaten
Titel
A study in facial regions saliency: a fuzzy measure approach
verfasst von
Paweł Karczmarek
Witold Pedrycz
Marek Reformat
Elaheh Akhoundi
Publikationsdatum
01.02.2014
Verlag
Springer Berlin Heidelberg
Erschienen in
Soft Computing / Ausgabe 2/2014
Print ISSN: 1432-7643
Elektronische ISSN: 1433-7479
DOI
https://doi.org/10.1007/s00500-013-1064-0

Weitere Artikel der Ausgabe 2/2014

Soft Computing 2/2014 Zur Ausgabe