Skin Detection A Bayesian Network Approach
Skin Detection A Bayesian Network Approach
net/publication/224752307
CITATIONS READS
63 464
4 authors, including:
T. Gevers
University of Amsterdam
305 PUBLICATIONS 16,800 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by T. Gevers on 03 June 2014.
where the summation is over the space of possible structures
and )( +* is the probability of error of the best
classifier learned with structure .
We use Metropolis-Hastings sampling to generate sam- Figure 1. An example of skin detection.
ples from the inverse error measure, without having to ever We use the database of Jones and Rehg [10] consists of
compute it for all possible structures. For constructing the 3,475 images containing skin and 8,796 non-skin images.
Metropolis-Hastings sampling, we define a neighborhood of Each image was manually segmented such that the skin pix-
a structure as the set of directed acyclic graphs to which we els are labeled. Examples of detected skin patches are pre-
can transit in the next step. Transition is done using a prede- sented in Figure 1. In the experiments we randomly se-
fined set of possible changes to the structure; at each transi- lected 3x3 skin and non-skin patches (100,000 in total). We
tion a change consists of a single edge addition, removal, or leave out 40,000 patches for testing and train the Bayesian
reversal. We define the acceptance probability of a candidate Network classifiers on the remaining 60,000. To compare
structure, b
, to replace a previous structure, 10 as follows:
-,/. <
b the results of the classifiers, we use the receiving operat-
:9 ;
ACB :<D =9 ;
IJIJKLI ANM < ing characteristic (ROC) curves. The ROC curves show, un-
243%576
=<>
@? B :9 E; D =<
F
243%5G6 9 ;
IJIJKLI ? M 9 ; F (5) der different classification thresholds, ranging from R to ,
'8 '8H
the probability of detecting a skin patch in a skin image,
YX[Z-\ ]X[Z-\ , against the probabil-
where O is the transition probability H from to , P
- -
: :
WV
is a temperature factor, and Q0 and are the sizes of
ity of falsely detecting a skin patch in a non-skin image,
the neighborhoods of 10 and
,/.
-,L.
respectively; this choice -
W^ V - 1_ % M
X[Z-\
`* 1_ % M
X[Z-\
.
corresponds to equal probability of transition to each mem- We first learn using all the training data being labeled (that
ber in the neighborhood of a structure. This further creates is 60,000 labeled patches). Figure 2 (left) shows the resul-
a Markov chain which is aperiodic and irreducible, thus sat- tant ROC curve for this case. The classifier learned with the
isfying the Markov chain Monte Carlo (MCMC) conditions. SSS algorithm outperforms both TAN and NB classifiers, and
Roughly speaking, P close to would allow acceptance of all perform quite well, achieving high detection rates with a
more structures with higher probability of error than previous low rate of false alarm. Next we remove the labels of some
structures. P close to R mostly allows acceptance of struc- of the training data and train the classifiers. Figure 2 (right)
tures that improve probability of error. A fixed P amounts to shows the case where the labels of 90% of the training data
1 1
0.9 0.9 satisfactory, then SSS can be used to attempt to further im-
0.8
0.7
0.8
0.7
prove performance. If none of the methods using the un-
labeled data improve performance over the supervised TAN
Detection
Detection
0.6 0.6
0.5 0.5
0.1
SSS
0.1
data using the active learning methodology. Of course, active
0
0 0.1 0.2 0.3 0.4 0.5
False Detection
0.6 0.7 0.8 0.9 1
0
0 0.1 0.2 0.3 0.4 0.5 0.6
False Detection
0.7 0.8 0.9 1 learning can be used as long as there are resources to label
some samples.
Figure 2. ROC curves showing detection rates of skin
Structure learning of Bayesian networks is not a topic mo-
compared to false detection with all data labeled (left) and
tivated solely by the use of unlabeled data. Skin detection
unlabeled data (right): SSS, NB learned with labeled
data only (NB-L) and with labeled and unlabeled data (NB- could be solved using classifiers other than Bayesian net-
LUL), and TAN learned with labeled data only (TAN-L) works. However, this work should be viewed as a combina-
and with labeled and unlabeled data (TAN-LUL). tion of 3 components; (1) the theory showing the limitations
of unlabeled data is used to motivate (2) the design of algo-
(leaving only 600 labeled patches) were removed. We see rithms to search for better performing structures of Bayesian
that the NB classifier using both labeled and unlabeled data networks and finally, (3) the successful application to skin de-
(NB-LUL) performs very poorly. The TAN based only on tection by learning with labeled and unlabeled data.
the 600 labeled images (TAN-L) and the TAN based on the
labeled and unlabeled images (TAN-LUL) are close in per- References
formance, thus there was no significant degradation of per- [1] S. Baluja. Probabilistic modeling for face orientation dis-
formance when adding the unlabeled data. crimination: Learning from labeled and unlabeled data. In
In Table 1 we summarize the results obtained for different NIPS, pages 854–860, 1998.
algorithms and in the presence of increasing number of un- [2] A. Bosson, G. Cawley, Y. Chan, and R. Harvey. Non-
labeled data. We fixed the false alarm to 1%, 5%, and 10% retrieval: blocking pornographic images. In CIVR, pages
and we computed the detection rates. Note that the detection 50–60, 2002.
[3] V. Castelli. The relative value of labeled and unlabeled sam-
rates for NB are lower than the ones obtained for the other ples in pattern recognition. PhD thesis, Stanford, 1994.
detectors. Overall, the results obtained with SSS are the best. [4] I. Cohen, F. Cozman, N. Sebe, M. Cirello, and T. Huang.
We see that even in the most difficult cases, there was suf- Semi-supervised learning of classifiers: Theory, algorithms,
ficient amount of unlabeled data to achieve almost the same and applications to human-computer interaction. IEEE
performance as with a large sized labeled dataset. Trans. on PAMI, to appear, 2004.
[5] T. Cormen, C. Leiserson, and R. Rivest. Introduction to al-
Table
1. Detection rates (%) for different false positives
gorithms. MIT Press, Cambridge, MA, 1990.
False
detections [6] M. Fleck, D. Forsyth, and C. Bregler. Finding naked people.
1% 5% 10%
Detector In ECCV, volume 2, pages 593–602, 1996.
[7] N. Friedman, D. Geiger, and M. Goldszmidt. Bayesian net-
60,000 labeled 64.74 85.26 90.64
work classifiers. Machine Learning, 29(2):131–163, 1997.
NB 600 labeled 61.34 81.27 84.85 [8] R. Ghani. Combining labeled and unlabeled data for multi-
600 labeled + 54,000 unlabeled 60.05 80.77 83.98 class text categorization. In ICML, pages 187–194, 2002.
60,000 labeled 86.85 96.22 99.00 [9] B. Jedynak, H. Zheng, and M. Daoudi. Statistical models for
600 labeled 84.50 88.84 93.63 skin detection. In CVPR Workshop on Statistical Analysis in
TAN
600 labeled + 54,000 unlabeled 84.66 88.82 93.01 Computer Vision, 2003.
[10] M. Jones and J. Rehg. Statistical color models with applica-
60,000 labeled 88.25 97.61 99.40 tion to skin detection. IJCV, 46(1):81–96, 2002.
SSS 600 labeled 85.23 92.51 96.15 [11] S. McKenna, S.Gong, and Y.Raja. Modelling facial colour
600 labeled + 54,000 unlabeled 87.66 95.82 98.32 and identity with gaussian mixtures. Pattern Recognition,
31:1883–1892, 1998.
5 Summary and Discussion [12] M. Meila. Learning with mixture of trees. PhD thesis, MIT,
1999.
In this work we presented a Bayesian Network approach [13] D. Saxe and R. Foulds. Toward robust skin identification in
for skin detection. We considered several instances of video images. In Automatic Face and Gesture Recognition,
Bayesian Networks and we suggested a methodology to per- pages 379–384, 1996.
form skin detection using both labeled and unlabeled data. [14] K. Schwerdt and J. Crowley. Robust face tracking using
In a nutshell, when faced with the option of learning with color. In Automatic Face and Gesture Recognition, pages
labeled and unlabeled data for skin detection using Bayesian 90–95, 2000.
[15] M.-H. Yang and N. Ahuja. Detecting human faces in color
networks, our discussion suggests using the following path. images. In ICIP, pages 127–130, 1998.
Start with Naive Bayes and TAN classifiers, learn only with [16] M.-H. Yang, D. Kriegman, and N. Ahuja. Detecting faces in
the available labeled data, and test whether the model is cor- images: A survey. PAMI, 24(1):34–58, 2002.
rect by learning with the unlabeled data. If the result is not