3.1 Feature Selection
3.1 Feature Selection
dimensionality
reduction
End w2
w 2
w
For a set S with N samples and d features
Initialize : w=(1,1,1….1)
For t= 1: T ( number iteration) do
Pick a random sample x from S
Find K nearmisses( x) y1, y 2 ,... y K
K nearhit( x) z1, z 2 ,...z K
( xi yi )2
with distance based on w, w ( x, y ) wi
2 2
i ( xi yi )
1 K 2 1 K 2
Compute : Dmiss
K j 1
( x, yi ) ; Dhit ( x, zij )
j
K j 1
Compute :
1
Dmiss Dhit
2
End
w2
w 2
w
29
Simba in semi-supervised contexte
Unknown labels Labels
Simba
[1] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Schölkopf, “Learning with local and
global consistency,” in Advances in Neural Information Processing Systems 16, 2004, pp.
321–328.
Unknown labels Labels
Simba
[2] F. Dornaika, Y. El Traboulsi, and A. Assoum, “Adaptive Two Phase Sparse Representation
Classifier for Face Recognition,” in Advanced Concepts for Intelligent Vision Systems: 15th
International Conference, ACIVS 2013, Poznań, Poland, October 28-31, 2013. Proceedings, J.
Blanc-Talon, A. Kasinski, W. Philips, D. Popescu, and P. Scheunders, Eds. Cham: Springer
International Publishing, 2013, pp. 182–191.
Must-link (x+y)
Cannot-link (x-y)
Compute nearhit and nearmiss
w w Fisher
33
The resulted subsets of many models of FS are strongly
dependent on the training set size.
It is not true that a large dimensionality input can always be
reduced to a small subset of features because the objective
feature is actually related with many input features and the
removal of any of them will seriously effect the learning
performance.
.
[1] F. Dornaika, Y. El Traboulsi, and A. Assoum, “Adaptive Two Phase Sparse Representation Classifier for
Face Recognition,” in Advanced Concepts for Intelligent Vision Systems: 15th International Conference,
ACIVS 2013, Poznań, Poland, October 28-31, 2013. Proceedings, J. Blanc-Talon, A. Kasinski, W. Philips, D.
Popescu, and P. Scheunders, Eds. Cham: Springer International Publishing, 2013, pp. 182–191.
[2] M. Yang, F. Wang, and P. Yang, “A novel feature selection algorithm based on hypothesis-margin,”
Journal of Computers, vol. 3, no. 12, pp. 27–34, 2008.
[3] K. Q. Weinberger and L. K. Saul, “Distance metric learning for large margin nearest neighbor
classification,” The Journal of Machine Learning Research, vol. 10, pp. 207–244, 2009.
[4] A. Moujahid, A. Abanda, and F. Dornaika, “Feature Extraction Using Block-based Local Binary Pattern
for Face Recognition,” Electronic Imaging, vol. 2016, no. 10, pp. 1–6, 2016.
[5] Y. Li and B.-L. Lu, “Feature selection based on loss-margin of nearest neighbor classification,” Pattern
Recognition, vol. 42, no. 9, pp. 1914–1921, Sep. 2009.
[6] W. Pan, P. Ma, and X. Su, “Feature Weighting Algorithm Based on Margin and Linear Programming,” in
Rough Sets and Current Trends in Computing, 2012, pp. 388–396.
[7] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Schölkopf, “Learning with local and global
consistency,” in Advances in Neural Information Processing Systems 16, 2004, pp. 321–328.
[8] K. Crammer, R. Gilad-Bachrach, A. Navot, and N. Tishby, “Margin analysis of the LVQ algorithm,” in
Advances in neural information processing systems, 2002, pp. 462–469.
[9]R. Gilad-Bachrach, A. Navot, and N. Tishby, “Margin based feature selection-theory and algorithms,” in
Proceedings of the twenty-first international conference on Machine learning, 2004, p. 43.
35