Algorithms 20130703 PDF
Algorithms 20130703 PDF
Algorithms 20130703 PDF
tool (2013-07-03). This list is grouped into different families and it is summarised in the following table.
Evolutionary Postprocessing FRBS: Selection and Tuning (14) Neural Networks for Regression (8) Neural Networks for Evolutionary Neural Networks for Regression (10) Regression (2) Support Vector Machines for Regression (2) Evolutionary Fuzzy Symbolic Regression (4) Statistical Regression (2) Resampling Data Over-sampling Methods (12) Space (20) Under-sampling Methods (8)
Cost-Sensitive Classification (3) Ensembles for Class Imbalance (19) Subgroup Discovery (7) Multi Instance Learning (9) Clustering Algorithms (1) Association Rules (11) Test Analysis (12) Post-Hoc Procedures for 1 x N Tests Statistical Tests (8) Post-Hoc Procedures (24) (12) Post-Hoc Procedures for N x N Tests (4)
Data Preprocessing
DISCRETIZATION
Full Name
Uniform Width Discretizer
Short Name
UniformWidth-D
Reference
H. Liu, F. Hussain, C.L. Tan, M. Dash. Discretization: An Enabling Technique. Data Mining and Knowledge Discovery 6:4 (2002) 393-423. H. Liu, F. Hussain, L. Tan, M. Dash. Discretization: An Enabling Technique. Data Mining and Knowledge Discovery 6:4 (2002) 393-423. U.M. Fayyad, K.B. Irani. Multi-Interval Discretization of Continuous-Valued Attributes for Classification Learning. 13th International Joint Conference on Uncertainly in Artificial Intelligence (IJCAI93). Chambery (France, 1993) 1022-1029. J.R. Quinlan. Induction of Decision Trees. Machine Learning 1 (1986) 81-106. X. Wu. A Bayesian Discretizer for Real-Valued Attributes. The. Computer J. 39:8 (1996) 688691. J. Cerquides, R. Lpez de Mntaras. Proposal and Empirical Comparison of a Parallelizable DistanceBased Discretization Method. 3rd International Conference on Knowledge Discovery and Data Mining (KDD99). NewPort Beach (USA, 1999) 139-142. R. Girldez, J.S. Aguilar-Ruiz, J.C. Riquelme, F. Ferrer-Troyano, D. Rodrguez. Discretization Oriented to Decision Rules Generation. In: L.C. Jain, E. Damiani, R.J. Howlett, N. Ichalkaranje (Eds.) Frontiers in Artificial Intelligence and Applications 82, 2002, 275-279. R. Girldez, J.S. Aguilar-Ruiz, J.C. Riquelme. Discretizacion Supervisada no Paramtrica Orientada a la Obtencion de Reglas de Decision.. IX Conferencia de la Asociacin Espaola de Inteligencia Artificial (CAEPIA'01). Gijn (Espaa, 2001) 53-62.
UniformFrequency-D
Fayyad-D
ID3-D
Bayesian-D
MantarasDist-D
USD-D
Chi-Merge Discretizer
ChiMerge-D
R. Kerber. ChiMerge: Attributes. National Intelligence American Intelligence (AAAI'92). 1992) 123-128.
Discretization of Numeric Conference on Artifical Association for Artificial San Jos (California USA,
Chi2 Discretizer
Chi2-D
H. Liu, R. Setiono. Feature Selection via Discretization. IEEE Transactions on Knowledge and Data Engineering 9:4 (1997) 642-645.
L. Gonzalez-Abril, F.J. Cuberos, F. Velasco, J.A. Ortega. Ameva: An autonomous discretization algorithm. Expert Systems with Applications 36 (2009) 5327-5332. K.M. Ho, P.D. Scott. Zeta: A Global Method for Discretization of Cotitinuous Variables. 3rd International Conference on Knowledge Discovery and Data Mining (KDD99). NewPort Beach (USA, 1999) 191-194. J.Y. Ching, A.K.C. W ong, K.C.C. Chan. ClassDependent Discretization for Inductive Learning from Continuous and Mixed-Mode Data. IEEE Transactions on Pattern Analysis and Machine Intelligence 17:7 (1995) 641-651. L.A. Kurgan, K.J. Cios. CAIM Discretization Algorithm. IEEE Transactions on Knowledge and Data Engineering 16:2 (2004) 145-153. C.-T. Sun, J.H. Hsu. An Extended Chi2 Algorithm for Discretization of Real Value Attributes. IEEE Transactions on Knowledge and Data Engineering 17:3 (2005) 437-441. Y. Yang, G.I. Webb. Discretization for naive-Bayes learning: managing discretization bias and variance. Machine Learning 74 (2009) 39-74. M. Boulle. Khiops: A Statistical Discretization Method of Continuous Attributes. Machine Learning 55:1 (2004) 53-69. F.E.H. Tay, L. Shen. A Modified Chi2 Algorithm for Discretization. IEEE Transactions on Knowledge and Data Engineering 14:2 (2002) 666-670. M. Boulle. MODL: A bayes optimal discretization method for continuous attributes. Machine Learning 65:1 (2006) 131-165. R.C. Holte. Very simple classification rules perform well on most commonly used datasets. Machine Learning 11 (1993) 63-91. Y. Yang, G.I. Webb. Discretization for naive-Bayes learning: managing discretization bias and variance. Machine Learning 74 (2009) 39-74. X. Liu. A Discretization Algorithm Based on a Heterogeneity Criterion. IEEE Transactions on Knowledge and Data Engineering 17:9 (2005) 1166-1173. C. Lee. A Hellinger-based discretization method for numeric attributes in classification learning. Knowledge-Based Systems 20:4 (2007) 419-425. Q.X. W u, D.A. Bell, G. Prasad, T.M. McGinnity. A Distribution-Index-Based Discretizer for DecisionMaking with Symbolic AI Approaches. IEEE Transactions on Knowledge and Data Engineering 19:1 (2007) 17-28.
Zeta Discretizer
Zeta-D
CADD-D
CAIM-D
ExtendedChi2-D
FixedFrequency-D
ModifiedChi2-D
MODL-D
1R Discretizer
1R-D
Proportional Discretizer
Proportional-D
Discretization HeterDisc-D Algorithm Based on a Heterogeneity Criterion Hellinger-based Discretizer DistributionIndex-Based Discretizer HellingerBD-D
DIBD-D
UCPD-D
S. Mehta, S. Parthasarathy, H. Yang. Toward Unsupervised Correlation Preserving Discretization. IEEE Transactions on Knowledge and Data Engineering 17:9 (2005) 1174-1185. F.J. Ruiz, C. Angulo, N. Agell. IDD: A Supervised Interval Distance-Based Method for Discretization. IEEE Transactions on Knowledge and Data Engineering 20:9 (2008) 1230-1238. C.J. Tsai, C.-I. Lee, W.-P. Yang. A discretization algorithm based on Class-Attribute Contingency Coefficient. Information Sciences 178:3 (2008) 714-731. P. Yang, J.-S. Li, Y.-X. Huang. HDD: a hypercube division-based algorithm for discretisation. International Journal of Systems Science 42:4 (2011) 557-566. M.R. Chmielewski, J.W. Grzymala-Busse. Global discretization of continuous attributes as preprocessing for Machine Learning. International Journal of Approximate Reasoning 15 (1996) 319331. S.D. Bay. Multivariate Discretization for Set Mining. Knowledge and Information Systems 3 (2001) 491-512. D.A. Zighed, R. Rabaseda, R. Rakotomalala. FUSINTER: A method for discretization of continuous attributes. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 6:3 (1998) 307-326.
Interval Distance- IDD-D Based Method for Discretization Discretization CACC-D algorithm based on Class-Attribute Contingency Coefficient Hypercube Division-Based HDD-D
Cluster Analysis
ClusterAnalysis-D
MVD-D
FUSINTER-D
FEATURE SELECTION
Full Name
Mutual Information Feature Selection Las Vegas Filter
Short Name
MIFS-FS
Reference
R. Battiti. Using Mutual Information For Selection Features In Supervised Neural Net Learning. IEEE Transactions on Neural Networks 5:4 (1994) 537550. H. Liu, R. Setiono. A Probabilistic Approach to Feature Selection: A Filter Solution. 13th International Conference on Machine Learning (ICML96 ). Bari (Italy, 1996) 319-327. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
LVF-FS
FOCUS
Focus-FS
H. Almuallim, T. Dietterich. Learning With Many Irrelevant Features. 9th National Conference on Artificial Intelligence (AAAI'91). Anaheim (California USA, 1991) 547-552.
Relief
Relief-FS
K. Kira, L. Rendell. A Practical Approach to Feature Selection. 9th International Workshop on Machine Learning (ML'92). Aberdeen (Scotlant UK, 1992) 249-256. H. Liu, R. Setiono. Feature Selection and Classification: A Probabilistic Wrapper Approach. 9th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems (IEA-AIE'96). Fukuoka (Japon, 1996) 419-424. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
LVW-FS
Automatic Branch ABB-IEP-FS and Bound using Inconsistent Examples Pairs Measure
H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
Full-LIU-FS
H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
Full-MI-FS
H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502.
H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. Relief-F Relief-F-FS I. Kononenko. Estimating Attributes: Analysis and Extensions of RELIEF. European Conference on Machine Learning 1994 (ECML94). Catania (Italy, 1994) 171-182. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. Simulated Annealing using Inconsistent Examples Pairs measure SA-IEP-FS H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. SA-LIU-FS H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. SA-MI-FS H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
SBS-MI-FS
H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998. H. Liu, L. Yu. Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Transactions on Knowledge and Data Engineering 17:4 (2005) 491-502. H. Liu, H. Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.
P.L. Lanzi. Fast Feature Selection With Genetic Algorithms: A Filter Approach. IEEE International Conference on Evolutionary Computation. Indianapolis. Indianapolis (USA, 1997) 537-540.
J. Yang, V. Honavar. Feature Subset Selection Using a Genetic Algorithm. IEEE Intelligent Systems 13:2 (1998) 44-49.
Short Name
POP-TSS
Reference
J.C. Riquelme, J.S. Aguilar-Ruiz, M. Toro. Finding representative patterns with ordered projections. Pattern Recognition 36 (2003) 1009-1018. M. Sebban, R. Nock, S. Lallich. Stopping Criterion for Boosting-Based Data Reduction Techniques: from Binary to Multiclass Problems. Journal of Machine Learning Research 3 (2002) 863-885. D.G. Lowe. Similarity Metric Learning For A Variable-Kernel Classifier. Neural Computation 7:1 (1995) 72-85. D.L. Wilson. Asymptotic Properties Of Nearest Neighbor Rules Using Edited Data. IEEE Transactions on Systems, Man and Cybernetics 2:3 (1972) 408-421. P.A. Devijver. On the editing rate of the MULTIEDIT algorithm. Pattern Recognition Letters 4:1 (1986) 9-12. J.S. Snchez, F. Pla, F.J. Ferri. Prototype selection for the nearest neighbor rule through proximity graphs. Pattern Recognition Letters 18 (1997) 507-513. K. Hattori, M. Takahashi. A new edited k-nearest neighbor rule in the pattern classification problem. Pattern Recognition 33 (2000) 521-528. J.S. Snchez, R. Barandela, A.I. Mrques, R. Alejo, J. Badenas. Analysis of new techniques to obtain quality training sets. Pattern Recognition Letters 24 (2003) 1015-1022. M. Grochowski, N. Jankowski. Comparison of instance selection algorithms I. Algorithms survey. VII International Conference on Artificial Intelligence and Soft Computing (ICAISC'04). LNCS 3070, Springer 2004, Zakopane (Poland, 2004) 598-603. F. Vazquez, J.S. Snchez, F. Pla. A stochastic approach to Wilson's editing algorithm. 2nd Iberian Conference on Pattern Recognition and Image Analysis (IbPRIA05). LNCS 3523, Springer 2005, Estoril (Portugal, 2005) 35-42. I. Tomek. An Experiment With The Edited NearestNeighbor Rule. IEEE Transactions on Systems, Man and Cybernetics 6:6 (1976) 448-452. C.E. Brodley. Adressing The Selective Superiority Problem: Automatic Algorithm/Model Class Selection. 10th International Machine Learning Conference (ICML'93). Amherst (MA USA, 1993) 17-24.
Prototipe PSRCG-TSS Selection by Relative Certainty Gain Variable Similarity VSM-TSS Metric Edited Nearest Neighbor ENN-TSS
Multiedit
Multiedit-TSS
Prototipe Selection based on Relative Neighbourhood Graphs Modified Edited Nearest Neighbor Nearest Centroid Neighbourhood Edition Edited NRBF
RNG-TSS
MENN-TSS
NCNEdit-TSS
ENRBF-TSS
ENNTh-TSS
AllKNN-TSS
ModelCS-TSS
Generational GGA-TSS Genetic Algorithm for Instance Selection Steady-State SGA-TSS Genetic Algorithm for Instance Selection Population-Based Incremental Learning PBIL-TSS
MISSING VALUES
Full Name
Delete Instances with Missing Values
Short Name
Ignore-MV
Reference
P.A. Gourraud, E. Ginin, A. Cambon-Thomsen. Handling Missing Values In Population Data: Consequences For Maximum Likelihood Estimation Of Haplotype Frequencies. European Journal of Human Genetics 12:10 (2004) 805-812. D.K.Y. Chiu, A.K.C. Wong. Synthesizing Knowledge: A Cluster Analysis Approach Using Event-Covering. IEEE Transactions on Systems, Man and Cybernetics, Part B 16:2 (1986) 251-259. G.E.A.P.A. Batista, M.C. Monard. An Analysis Of Four Missing Data Treatment Methods For Supervised learning. Applied Artificial Intelligence 17:5 (2003) 519-533. J.W. Grzymala-Busse, L.K. Goodwin, W.J. Grzymala-Busse, X. Zheng. Handling Missing Attribute Values in Preterm Birth Data Sets. 10th International Conference of Rough Sets, Fuzzy Sets, Data Mining and Granular Computing (RSFDGrC'05). LNCS 3642, Springer 2005, Regina (Canada, 2005) 342-351. J.W. Grzymala-Busse. On the Unknown Attribute Values In Learning From Examples. 6th International Symposium on Methodologies For Intelligent Systems (ISMIS91). Charlotte (USA, 1991) 368-377.
EventCovering-MV
KNN-MV
MostCommon-MV
AllPossible-MV
10
K-means Imputation
KMeans-MV
J. Deogun, W. Spaulding, B. Shuart, D. Li. Towards Missing Data Imputation: A Study of Fuzzy Kmeans Clustering Method. 4th International Conference of Rough Sets and Current Trends in Computing (RSCTC'04). LNCS 3066, Springer 2004, Uppsala (Sweden, 2004) 573-579. J.W. Grzymala-Busse, L.K. Goodwin, W.J. Grzymala-Busse, X. Zheng. Handling Missing Attribute Values in Preterm Birth Data Sets. 10th International Conference of Rough Sets, Fuzzy Sets, Data Mining and Granular Computing (RSFDGrC'05). LNCS 3642, Springer 2005, Regina (Canada, 2005) 342-351. J.W. Grzymala-Busse. On the Unknown Attribute Values In Learning From Examples. 6th International Symposium on Methodologies For Intelligent Systems (ISMIS91). Charlotte (USA, 1991) 368-377. J. Deogun, W. Spaulding, B. Shuart, D. Li. Towards Missing Data Imputation: A Study of Fuzzy Kmeans Clustering Method. 4th International Conference of Rough Sets and Current Trends in Computing (RSCTC'04). LNCS 3066, Springer 2004, Uppsala (Sweden, 2004) 573-579. H.A.B. Feng, G.C. Chen, C.D. Yin, B.B. Yang, Y.E. Chen. A SVM regression based approach to filling in Missing Values. 9th International Conference on Knowledge-Based and Intelligent Information and Engineering Systems (KES2005). LNCS 3683, Springer 2005, Melbourne (Australia, 2005) 581587. O. Troyanskaya, M. Cantor, G. Sherlock, P. Brown, T. Hastie, R. Tibshirani, D. Botstein, R.B. Altman. Missing value estimation methods for DNA microarrays. Bioinformatics 17 (2001) 520-525. S. Oba, M. Sato, I. Takemasa, M. Monden, K. Matsubara, S. Ishii. A Bayesian missing value estimation method for gene expression profile data. Bioinformatics 19 (2003) 2088-2096. T. Schneider. Analysis of incomplete climate data: Estimation of Mean Values and covariance matrices and imputation of Missing values. Journal of Climate 14 (2001) 853-871. H.A. Kim, G.H. Golub, H. Park. Missing value estimation for DNA microarray gene expression data: Local least squares imputation. Bioinformatics 21:2 (2005) 187-198. O. Troyanskaya, M. Cantor, G. Sherlock, P. Brown, T. Hastie, R. Tibshirani, D. Botstein, R.B. Altman. Missing value estimation methods for DNA microarrays. Bioinformatics 17 (2001) 520-525.
Assign All Posible Values of the Attribute Restricted to the Given Concept Fuzzy K-means Imputation
ConceptAllPossible-MV
FKMeans-MV
SVMimpute-MV
WKNNimpute-MV
Bayesian Principal BPCA-MV Component Analysis ExpectationMaximization single imputation Local Least Squares Imputation Single Vector Decomposition imputation EM-MV
LLSImpute-MV
SVDImpute-MV
11
TRANSFORMATION
Full Name
Decimal Scaling ranging Min Max ranging
Short Name
DecimalScaling-TR
Reference
L.A. Shalabi, Z. Shaaban, B. Kasasbeh. Data Mining: A Preprocessing Engine. Journal of Computer Science 2:9 (2006) 735-735. L.A. Shalabi, Z. Shaaban, B. Kasasbeh. Data Mining: A Preprocessing Engine. Journal of Computer Science 2:9 (2006) 735-735. L.A. Shalabi, Z. Shaaban, B. Kasasbeh. Data Mining: A Preprocessing Engine. Journal of Computer Science 2:9 (2006) 735-735. L.A. Shalabi, Z. Shaaban, B. Kasasbeh. Data Mining: A Preprocessing Engine. Journal of Computer Science 2:9 (2006) 735-735.
MinMax-TR
Z Score ranging
ZScore-TR
DATA COMPLEXITY
Full Name Short Name Reference
T.K. Ho, M. Basu. Complexity measures of supervised classification problems. IEEE Transactions on Pattern Analysis and Machine Intelligence 24:3 (2002) 289-300. Data Complexity Metrics-DC Metrics calculation
Short Name
SaturationFilter-F
Reference
D. Gamberger, N. Lavrac, S. Dzroski. Noise detection and elimination in data preprocessing: Experiments in medical domains. Applied Artificial Intelligence 14:2 (2000) 205-223. J.D. Hulse, T.M. Khoshgoftaar, H. Huang. The pairwise attribute noise detection algorithm. Knowledge and Information Systems 11:2 (2007) 171-190. D. Gamberger, N. Lavrac, C. Groselj. Experiments with noise filtering in a medical domain. 16th International Conference on Machine Learning (ICML99). San Francisco (USA, 1999) 143-151. X. Zeng, T. Martinez. A Noise Filtering Method Using Neural Networks. IEEE International Workshop on Soft Computing Techniques in Instrumentation, Measurement and Related Applications (SCIMA2003). Utah (USA, 2003) 2631. C.E. Brodley, M.A. Friedl. Identifying Mislabeled Training Data. Journal of Articial Intelligence Research 11 (1999) 131-167.
Pairwise Attribute PANDA-F Noise Detection Algorithm Filter Classification Filter ClassificationFilter-F
ANR-F
Ensemble Filter
EnsembleFilter-F
12
S. Verbaeten, A.V. Assche. Ensemble methods for noise elimination in classification problems. 4th International Workshop on Multiple Classier Systems (MCS 2003). LNCS 2709, Springer 2003, Guilford (UK, 2003) 317-325.
IterativePartitioning Filter
IterativePartitioningFilter- T.M. Khoshgoftaar, P. Rebours. Improving F software quality prediction by noise filtering techniques. Journal of Computer Science and Technology 22 (2007) 387-396.
13
Classification Algorithms
Short Name
AQ-C
Reference
R.S. Michalksi,, I. Mozetic, N. Lavrac. The Multipurpose Incremental Learning System AQ15 And Its Testing Application To Three Medical Domains. 5th INational Conference on Artificial Intelligence (AAAI'86 ). Philadelphia (Pennsylvania, 1986) 1041-1045. P. Clark, T. Niblett. The CN2 Induction Algorithm. Machine Learning Journal 3:4 (1989) 261-283. J. Cendrowska. PRISM: An algorithm for inducing modular rules. International Journal of ManMachine Studies 27:4 (1987) 349-370. R.C. Holte. Very simple classification rules perform well on most commonly used datasets. Machine Learning 11 (1993) 63-91. G. Gra, A. Wojna. RIONA: A New Classification System Combining Rule Induction and InstanceBased Learning. Fundamenta Informaticae 51:4 (2002) 1-22. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman Publishers, 1993. J.R. Quinlan. MDL and Categorical Theories (Continued). Machine Learning: Proceedings of the Twelfth International Conference. Lake Tahoe California (United States of America, 1995) 464470.
CN2 PRISM
CN2-C PRISM-C
1R
1R-C
Riona-C
C45Rules-C
J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman Publishers, 1993. J.R. Quinlan. MDL and Categorical Theories (Continued). Machine Learning: Proceedings of the Twelfth International Conference. Lake Tahoe California (United States of America, 1995) 464470. E. Frank, I.H. Witten. Generating Accurate Rule Sets Without Global Optimization. Proceedings of the Fifteenth International Conference on Machine Learning. (1998) 144-151. W.W. Cohen. Fast Effective Rule Induction. Machine Learning: Proceedings of the Twelfth International Conference. Lake Tahoe California (United States of America, 1995) 1-10. W.W. Cohen, Y. Singer. A Simple, Fast, and Effective Rule Learner. Proceedings of the Sixteenth National Conference on Artificial Intelligence. Orlando Florida (United States of America, 1999) 335-342. F. Berzal, J.C. Cubero, D. Snchez, J.M. Serrano.
PART
PART-C
Repeated Incremental Pruning to Produce Error Reduction Simple Learner with Iterative Pruning to Produce Error Reduction Association Rule
Ripper-C
Slipper-C
ART-C
14
Serrano.ART: A Hybrid Classification Model. Machine Learning 54 (2004) 67-92. L.A. Kurgan, K.J. Cios, S. Dick. Highly Scalable and Robust Rule Learner: Performance Evaluation and Comparison. IEEE Transactions on Systems, Man and Cybernetics,Part B: Cybernetics 36:1 (2006) 32-53. M. Sholom, N. Indurkhya. Optimized Rule Induction. IEEE Expert 1 (1993) 61-70. J. Stefanowski. On rough set based approaches to induction of decision rules. In: L. Polkowski, A. Skowron (Eds.) Rough sets in data mining and knowledge discovery, 1998, 500-529. J. Stefanowski. On rough set based approaches to induction of decision rules. In: L. Polkowski, A. Skowron (Eds.) Rough sets in data mining and knowledge discovery, 1998, 500-529. X. Wu, D. Urpani. Induction By Attribute Elimination. IEEE Transactions on Knowledge and Data Engineering 11:5 (1999) 805-812. D.T. Pham, A.A. Afify. RULES-6: A Simple Rule Induction Algorithm for Supporting Decision Making. 31st Annual Conference of IEEE Industrial Electronics Society (IECON). (2005) 2184-2189. D.T. Pham, A.A. Afify. SRI: a scalable rule induction algorithm. S.J. Hong. R-MINI: An Iterative Approach for Generating Minimal Rules from Examples. IEEE Transactions on Knowledge and Data Engineering 9:5 (1997) 709-717.
Swap1 Learning Examples Module 1 Learning Examples Module 2 Rule Induction Two In One Rule Extraction System Version 6
Swap1-C LEM1-C
LEM2-C
Ritio-C
Rules6-C
SRI-C RMini-C
15
Algorithm
Attributes Based Concepts. 6th European Conference on Machine Learning (ECML'93). Lecture Notes in Artificial Intelligence. Viena (Austria, 1993) 280-296. XCS-C Hider-C S.W. Wilson. Classifier Fitness Based on Accuracy. Evolutionary Computation 3:2 (1995) 149-175. J.S. Aguilar-Ruiz, J.C. Riquelme, M. Toro. Evolutionary learning of hierarchical decision rules.. Transactions on Systems, Man, and Cybernetics - Part B: Cybernetics 33:2 (2003) 324331. J.S. Aguilar-Ruiz, R. Girldez, J.C. Riquelme. Natural Encoding for Evolutionary Supervised Learning. IEEE Transactions on Evolutionary Computation 11:4 (2007) 466-479.
J. Bacardit, J.M. Garrell. Bloat control and generalization pressure using the minimum description length principle for a pittsburgh approach learning classifier system. Advances at the frontier of Learning Classifier Systems. Springer Berlin-Heidelberg. (2007) 61-80. M.L. Wong, K.S. Leung. Data Mining using grammar based genetic programming and applications. Kluwer Academics Publishers, 2000. E. Bernad-Mansilla, J.M. Garrell. Accuracy-Based Learning Classifier Systems: Models, Analysis and Applications to Classification Tasks. Evolutionary Computation 11:3 (2003) 209-238. T. Sousa, A. Silva, A. Neves. Particle Swarm based Data Mining Algorithms for classification tasks. Parallel Computing 30 (2004) 767-783.
LogenPro-C
UCS-C
Particle Swarm PSO_ACO-C Optimization / Ant Colony Optimization for Classification Ant Miner Ant_Miner-C
R.S. Parpinelli, H.S. Lopes, A.A. Freitas. Data Mining With an Ant Colony Optimization Algorithm. IEEE Transactions on Evolutionary Computation 6:4 (2002) 321-332.
Advanced_Ant_Miner-C R.S. Parpinelli, H.S. Lopes, A.A. Freitas. Data Mining With an Ant Colony Optimization Algorithm. IEEE Transactions on Evolutionary Computation 6:4 (2002) 321-332. R.S. Parpinelli, H.S. Lopes, A.A. Freitas. An Ant Colony Algorithm for Classification Rule Discovery. In: H.A. Abbass, R.A. Sarker, C.S. Newton (Eds.) Data Mining: a Heuristic Approach, 2002, 191-208.
Ant Miner+
Ant_Miner_Plus-C
R.S. Parpinelli, H.S. Lopes, A.A. Freitas. Data Mining With an Ant Colony Optimization Algorithm. IEEE Transactions on Evolutionary Computation 6:4 (2002) 321-332.
Advanced_Ant_Miner_Pl R.S. Parpinelli, H.S. Lopes, A.A. Freitas. Data us-C Mining With an Ant Colony Optimization Algorithm.
16
IEEE Transactions on Evolutionary Computation 6:4 (2002) 321-332. R.S. Parpinelli, H.S. Lopes, A.A. Freitas. An Ant Colony Algorithm for Classification Rule Discovery. In: H.A. Abbass, R.A. Sarker, C.S. Newton (Eds.) Data Mining: a Heuristic Approach, 2002, 191-208. Constricted Particle Swarm Optimization CPSO-C T. Sousa, A. Silva, A. Neves. Particle Swarm based Data Mining Algorithms for classification tasks. Parallel Computing 30 (2004) 767-783. T. Sousa, A. Silva, A. Neves. Particle Swarm based Data Mining Algorithms for classification tasks. Parallel Computing 30 (2004) 767-783. Y. Liu, Z. Qin, Z. Shi, J. Chen. Rule Discovery with Particle Swarm Optimization. Advanced Workshop on Content Computing (AWCC). LNCS 3309, Springer 2004 (2004) 291-296. J. Bacardit, E. Burke, N. Krasnogor. Improving the scalability of rule-based evolutionary learning. Memetic computing 1:1 (2009) 55-67.
Linear Decreasing LDWPSO-C Weight - Particle Swarm Optimization Real Encoding Particle Swarm Optimization Bioinformaticsoriented hierarchical evolutionary learning REPSO-C
BioHel-C
D.P. Greene, S.F. Smith. Competitionbased induction of decision models from examples. Machine Learning 13:23 (1993) 229-257. K.C. Tan, Q. Yu, J.H. Ang. A coevolutionary algorithm for rules discovery in data mining. International Journal of Systems Science 37:12 (2006) 835-864. W.H. Au, K.C.C. Chan, X. Yao. A novel evolutionary data mining algorithm with applications to churn prediction. IEEE Transactions on Evolutionary Computation 7:6 (2003) 532-545. C.Z. Janikow. A knowledgeintensive genetic algorithm for supervised learning. Machine Learning 13:2 (1993) 189-228. L. Jiao, J. Liu, W. Zhong. An organizational coevolutionary algorithm for classification. IEEE Transactions on Evolutionary Computation 10:12 (2006) 67-80. F. Zhu, S.U. Guan. Ordered incremental training with genetic algorithms. International Journal of Intelligent Systems 19:12 (2004) 1239-1256. S.U. Guan, F. Zhu. An incremental approach to genetic algorithmsbased classification. IEEE Transactions on Systems, Man, and Cybernetics, Part B 35:2 (2005) 227-239. J. Bacardit, N. Krasnogor. Performance and Efficiency of Memetic Pittsburgh Learning Classifier Systems. Evolutionary Computation 17:3 (2009) 307-342.
DMEL-C
Genetic-based GIL-C Inductive Learning Organizational Co- OCEC-C Evolutionary algorithm for Classification Ordered OIGA-C Incremental Genetic Algorithm Incremental Learning with Genetic Algorithms ILGA-C
17
C.C. Bojarczuk, H.S. Lopes, A.A. Freitas, E.L. Michalkiewicz. A constrained-syntax genetic programming system for discovering classification rules: applications to medical datasets. Artificial Intelligence in Medicine 30:1 (2004) 27-48. I.D. Falco, A.D. Cioppa, E. Tarantino. Discovering interesting classification rules with genetic programming. Applied Soft Computing 1 (2002) 257-269. K.C. Tan, A. Tay, T.H. Lee, C.M. Heng. Mining multiple comprehensible classification rules using genetic programming. he 2002 Congress on Evolutionary Computation (CEC02). Piscataway (USA, 2002) 1302-1307. A. Pietramala, V.L. Policicchio, P. Rullo, I. Sidhu. A Genetic Algorithm for Text Classification Rule Induction. European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2008). LNCS 5212, Springer 2008, Antwerp (Belgium, 2008) 188-203.
Falco_GP-C
Tan_GP-C
Genetic Algorithm Olex-GA-C designed for the task of solving problem MAX-F
PDFC-C
18
Short Name
GFS-SP-C
Reference
L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators With SA Search To Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-192.
GFS-GPG-C
L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators With SA Search To Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-192. M.J. del Jesus, F. Hoffmann, L. Junco, L. Snchez. Induction of Fuzzy-Rule-Based Classifiers With Evolutionary Boosting Algorithms. IEEE Transactions on Fuzzy Systems 12:3 (2004) 296308. J. Otero, L. Snchez. Induction of Descriptive Fuzzy Classifiers With The Logitboost Algorithm. Soft Computing 10:9 (2006) 825-835. L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators With SA Search To Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-192. L. Snchez, J. Otero. Boosting Fuzzy Rules in Classification Problems Under Single-Winner Inference. International Journal of Intelligent Systems 22:9 (2007) 1021-1034. H. Ishibuchi, K. Nozaki, N. Yamamoto, H. Tanaka. Selecting Fuzzy If-Then Rules for Classification. IEEE Transactions on Fuzzy Systems 3:3 (1995) 260-270. A. Gonzlez, R. Perez. Selection of relevant features in a fuzzy genetic learning algorithm. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 31:3 (2001) 417-425.
GFS-AdaBoost-C
LogitBoost
GFS-LogitBoost-C
Fuzzy Learning based on Genetic Programming Logitboost with Single Winner Inference Grid Rule Base Generation and Genetic Rule Selection
GFS-GP-C
GFS-MaxLogitBoost-C
GFS-Selec-C
Structural SLAVE-C Learning Algorithm in a Vague Environment with Feature Selection Methodology to MOGUL-C Obtain Genetic fuzzy rule-based systems Under the iterative Learning approach
O. Cordn, M.J. del Jesus, F. Herrera. Genetic learning of fuzzy rule-based classification systems cooperating with fuzzy reasoning methods. International Journal of Intelligent Systems 13:10 (1998) 1025-1053. O. Cordn, M.J. del Jesus, F. Herrera, M. Lozano. MOGUL: A Methodology to Obtain Genetic fuzzy rule-based systems Under the iterative rule Learning approach. International Journal of Intelligent Systems 14:11 (1999) 1123-1153. H. Ishibuchi, T. Nakashima, T. Murata. Performance evaluation of fuzzy classifier systems for
GFS-GCCL-C
19
multidimensional pattern classification problems. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 29:5 (1999) 601618. H. Ishibuchi, T. Yamamoto, T. Nakashima. Hybridization of Fuzzy GBML Approaches for Pattern Classification Problems. IEEE Transactions on Systems, Man and Cybernetics - Part B: Cybernetics 35:2 (2005) 359-365. Y. Shi, R. Eberhart, Y. Chen. Implementation of evolutionary fuzzy systems. IEEE Transactions on Fuzzy Systems 7:2 (1999) 109-119. E.G. Mansoori, M.J. Zolghadri, S.D. Katebi. SGERD: A Steady-State Genetic Algorithm for Extracting Fuzzy Classification Rules From Data. IEEE Transactions on Fuzzy Systems 16:4 (2008) 10611071.
GFS-ES-C
Steady-State SGERD-C Genetic Algorithm for Extracting Fuzzy Classification Rules From Data
Short Name
BNGE-C
Reference
D. Wettschereck, T.G. Dietterich. An Experimental Comparison of the Nearest-Neighbor and NearestHyperrectangle Algorithms. Machine Learning 19 (1995) 5-27. S. Salzberg. A Nearest Hyperrectangle Learning Method. Machine Learning 6 (1991) 251-276. O. Luaces. Inflating examples to obtain rules. International Journal of Intelligent Systems 18 (2003) 1113-1143. P. Domingos. Unifying Instance-Based and RuleBased Induction. Machine Learning 24:2 (1996) 141-168.
EACH-C
INNER-C
RISE-C
ASSOCIATIVE CLASSIFICATION
Full Name
Classification Based on Associations Classification Based on Associations 2
Short Name
CBA-C
Reference
B. Liu, W. Hsu, Y. Ma. Integrating Classification and Association Rule Mining. 4th International Conference on Knowledge Discovery and Data Mining (KDD98). New York (USA, 1998) 80-86. B. Liu, Y. Ma, C.K. Wong. Classification Using Association Rules: Weaknesses and Enhancements . In: R.L. Grossman, C. Kamath, V. Kumar (Eds.) Data Mining for Scientific and Engineering Applications, 2001, 591-601.
CBA2-C
20
Classification based on Predictive Association Rules Classification Based on Multiple Class-Association Rules
CPAR-C
X. Yin, J. Han. CPAR: Classification based on Predictive Association Rules. 3rd SIAM International Conference on Data Mining (SDM03). San Francisco (USA, 2003) 331-335. W. Li, J. Han, J. Pei. CMAR: Accurate and efficient classification based on multiple class-association rules. 2001 IEEE International Conference on Data Mining (ICDM01). San Jose (USA, 2001) 369-376. Y.-C. Hu, R.-S. Chen, G.-H. Tzeng. Finding fuzzy classification rules using data mining techniques. Pattern Recognition Letters 24:1-3 (2003) 509519. Z. Chen, G. Chen. Building an associative classifier based on fuzzy association rules. International Journal of Computational Intelligence Systems 1:3 (2008) 262-273. J. Alcala-Fdez, R. Alcal, F. Herrera. A Fuzzy Association Rule-Based Classification Model for High-Dimensional Problems with Genetic Rule Selection and Lateral Tuning. IEEE Transactions on Fuzzy Systems 19:5 (2011) 857-872.
CMAR-C
Fuzzy rules for FCRA-C classification problems based on the Apriori algorithm Classification with Fuzzy Association Rules Fuzzy Association Rule-based Classification method for HighDimensional problems CFAR-C
FARC-HD-C
DECISION TREES
Full Name
C4.5 Iterative Dicotomizer 3 Classification and Regression Tree
Short Name
C4.5-C ID3-C CART-C
Reference
J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. J.R. Quinlan. Induction of Decision Trees. Machine Learning 1 (1986) 81-106. L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone. Classification and Regression Trees. Chapman and Hall (Wadsworth, Inc.), 1984. M. Mehta, R. Agrawal, J. Rissanen. SLIQ: A Fast Scalable Classifier for Data Mining. Proceedings of the 5th International Conference on Extending Database Technology. (1996) 18-32. D.R. Carvalho, A.A. Freitas. A hybrid decision tree/genetic algorithm method for data mining. Information Sciences 163:1 (2004) 13-35. E. Cant-Paz, C. Kamath. Inducing oblique decision trees with evolutionary algorithms. IEEE Transactions on Evolutionary Computation 7:1 (2003) 54-68. J. Gama. Functional Trees. Machine Learning 55 (2004) 219-250. R. Rastogi, K. Shim. PUBLIC: A Decision Tree Classifier that Integrates Building and Pruning. Data Mining and Knowledge Discovery 4:4 (2000) 315-344.
Hybrid Decision Tree -Genetic Algorithm Oblique Decision Tree with Evolutionary Learning Functional Trees PrUning and BuiLding Integrated in Classification
DT_GA-C
DT_Oblique-C
FunctionalTrees-C PUBLIC-C
21
J.B. Gray, G. Fan. Classification tree analysis using TARGET. Computational Statistics and Data Analysis 52:3 (2008) 1362-1372.
PROTOTYPE SELECTION
Full Name
All-KNN
Short Name
AllKNN-C
Reference
I. Tomek. An Experiment With The Edited NearestNeighbor Rule. IEEE Transactions on Systems, Man and Cybernetics 6:6 (1976) 448-452. D.L. Wilson. Asymptotic Properties Of Nearest Neighbor Rules Using Edited Data. IEEE Transactions on Systems, Man and Cybernetics 2:3 (1972) 408-421. P.A. Devijver. On the editing rate of the MULTIEDIT algorithm. Pattern Recognition Letters 4:1 (1986) 9-12. P.E. Hart. The Condensed Nearest Neighbour Rule. IEEE Transactions on Information Theory 14:5 (1968) 515-516. I. Tomek. Two modifications of CNN. IEEE Transactions on Systems, Man and Cybernetics 6 (1976) 769-772. D.W. Aha, D. Kibler, M.K. Albert. Instance-Based Learning Algorithms. Machine Learning 6:1 (1991) 37-66. K. Hattori, M. Takahashi. A new edited k-nearest neighbor rule in the pattern classification problem. Pattern Recognition 33 (2000) 521-528. V.S. Devi, M.N. Murty. An incremental prototype set building technique. Pattern Recognition 35 (2002) 505-513. D.R. Wilson, T.R. Martinez. Reduction Tecniques For Instance-Based Learning Algorithms. Machine Learning 38:3 (2000) 257-286. H. Brighton, C. Mellish. Advances In Instance Selection For Instance-Based. Data mining and Knowledge Discovery 6:2 (2002) 153-172. G.W. Gates. The Reduced Nearest Neighbour Rule. IEEE Transactions on Information Theory 18:3 (1972) 431-433. G.L. Ritter, H.B. Woodruff, S.R. Lowry, T.L. Isenhour. An Algorithm For A Selective Nearest Neighbor Decision Rule. IEEE Transactions on Information Theory 21:6 (1975) 665-669. D.G. Lowe. Similarity Metric Learning For A Variable-Kernel Classifier. Neural Computation 7:1 (1995) 72-85.
ENN-C
Multiedit
Multiedit-C
Condensed Nearest Neighbor Tomek's modification of Condensed Nearest Neighbor Instance Based 3
CNN-C
TCNN-C
IB3-C
Modified Edited Nearest Neighbor Modified Condensed Nearest Neighbor Decremental Reduction Optimization Procedure 3 Iterative Case Filtering Reduced Nearest Neighbor Selective Nearest Neighbor
MENN-C
MCNN-C
DROP3-C
ICF-C
RNN-C
SNN-C
22
PSC-C
J.A. Olvera-Lpez, J.A. Carrasco-Ochoa, J.F. Martnez-Trinidad. A new fast prototype selection method based on clustering. Pattern Analysis and Applications 13 (2010) 131-141. C.E. Brodley. Adressing The Selective Superiority Problem: Automatic Algorithm/Model Class Selection. 10th International Machine Learning Conference (ICML'93). Amherst (MA USA, 1993) 17-24. D. Kibler, D.W. Aha. Learning Representative Exemplars Of Concepts: An Initial Case Study. 4th International Workshop on Machine Learning (ML'87 ). Irvine (CA USA, 1987) 24-30. B.V. Dasarathy. Minimal Consistent Set (MCS) Identification for Optimal Nearest Neighbor Decision Systems Design. IEEE Transactions on Systems, Man and Cybernetics 24:3 (1994) 511517. D.B. Skalak. Prototype and Feature Selection by Sampling and Random Mutation Hill Climbing Algorithms. 11th International Conference on Machine Learning (ML'94). New Brunswick (NJ USA, 1994) 293-301. X.-Z. Wang, B. Wu, Y.-L. He, X.-H. Pei. NRMCS: Noise removing based on the MCS. 7th International Conference on Machine Learning and Cybernetics (ICMLA08). La Jolla Village (USA, 2008) 89-93. E. Marchiori. Class Conditional Nearest Neighbor for Large Margin Instance Selection. IEEE Transactions on Pattern Analysis and Machine Intelligence 32:2 (2010) 364-370. K. Chidananda-Gowda, G. Krishna. The Condensed Nearest Neighbor Rule Using Concept of Mutual Nearest Neighborhood. IEEE Transactions on Information Theory 25:4 (1979) 488-490. R.M. Cameron-Jones. Instance selection by encoding length heuristic with random mutation hill climbing. 8th Australian Joint Conference on Artificial Intelligence (AJCAI-95). (Australia, 1995) 99-106. J.S. Snchez, F. Pla, F.J. Ferri. Prototype selection for the nearest neighbor rule through proximity graphs. Pattern Recognition Letters 18 (1997) 507513. J.S. Snchez, F. Pla, F.J. Ferri. Prototype selection for the nearest neighbor rule through proximity graphs. Pattern Recognition Letters 18 (1997) 507513. J.S. Snchez, R. Barandela, A.I. Mrques, R. Alejo, J. Badenas. Analysis of new techniques to obtain quality training sets. Pattern Recognition Letters 24
ModelCS-C
Shrink
Shrink-C
MCS-C
NRMCS-C
MNV-C
Explore-C
Prototipe GG-C Selection based on Gabriel Graphs Prototipe Selection based on Relative Neighbourhood Graphs Nearest Centroid Neighbourhood Edition RNG-C
NCNEdit-C
23
(2003) 1015-1022. Tabu Search for ZhangTS-C Instance Selection Prototipe PSRCG-C Selection by Relative Certainty Gain C-Pruner CPruner-C H. Zhang, G. Sun. Optimal reference subset selection for nearest neighbor classification by tabu search. Pattern Recognition 35 (2002) 1481-1490. M. Sebban, R. Nock, S. Lallich. Stopping Criterion for Boosting-Based Data Reduction Techniques: from Binary to Multiclass Problems. Journal of Machine Learning Research 3 (2002) 863-885. K.P Zhao, S.G. Zhou, J.H. Guan, A.Y. Zhou. CPruner: An improved instance prunning algorithm. Second International Conference on Machine Learning and Cybernetics (ICMLC'03). Xian (China, 2003) 94-99. J.C. Riquelme, J.S. Aguilar-Ruiz, M. Toro. Finding representative patterns with ordered projections. Pattern Recognition 36 (2003) 1009-1018. M.T. Lozano, J.S. Snchez, F. Pla. Using the geometrical distribution of prototypes for training set condesing. 10th Conference of the Spanish Association for Artificial Intelligence (CAEPIA03). LNCS 3040, Springer 2003, Malaga (Spain, 2003) 618-627. M. Grochowski, N. Jankowski. Comparison of instance selection algorithms I. Algorithms survey. VII International Conference on Artificial Intelligence and Soft Computing (ICAISC'04). LNCS 3070, Springer 2004, Zakopane (Poland, 2004) 598-603. R. Barandela, F.J. Ferri, J.S. Snchez. Decision boundary preserving prototype selection for nearest neighbor classification. International Journal of Pattern Recognition and Artificial Intelligence 19:6 (2005) 787-806. F. Vazquez, J.S. Snchez, F. Pla. A stochastic approach to Wilson's editing algorithm. 2nd Iberian Conference on Pattern Recognition and Image Analysis (IbPRIA05). LNCS 3523, Springer 2005, Estoril (Portugal, 2005) 35-42. Y. Li, Z. Hu, Y. Cai, W. Zhang. Support vector vased prototype selection method for nearest neighbor rules. I International conference on advances in natural computation (ICNC05). LNCS 3610, Springer 2005, Changsha (Chine, 2005) 528-535. J.A. Olvera-Lpez, J.F. Martnez-Trinidad, J.A. Carrasco-Ochoa. Edition Schemes Based on BSE. 10th Iberoamerican Congress on Pattern Recognition (CIARP2004). LNCS 3773, Springer 2005, La Havana (Cuba, 2005) 360-367. F. Angiulli. Fast nearest neighbor condensation for large data sets classification. IEEE Transactions on Knowledge and Data Engineering 19:11 (2007) 1450-1464.
POP-C
Reconsistent-C
Edited NRBF
ENRBF-C
Edited Nearest Neighbor with Estimation of Probabilities Threshold Support Vector based Prototype Selection
ENNTh-C
SVBPS-C
FCNN-C
24
HMNEI-C
E. Marchiori. Hit Miss Networks with Applications to Instance Selection. Journal of Machine Learning Research 9 (2008) 997-1017. F. Chang, C.C. Lin, C.-J. Lu. Adaptive Prototype Learning Algorithms: Theoretical and Experimental Studies. Journal of Machine Learning Research 7 (2006) 2125-2148.
GCNN-C
Generational GGA-C Genetic Algorithm for Instance Selection Steady-State SGA-C Genetic Algorithm for Instance Selection Cooperative CoCoIS-C Coevolutionary Instance Selection Population-Based Incremental Learning PBIL-C
Intelligent Genetic IGA-C Algorithm for Edition Genetic Algorithm GA_MSE_CC_FSM-C for Editing k-NN with MSE estimation, clustered crossover and fast smart mutation Steady-State SSMA-C Memetic Algorithm for Instance Selection
25
PROTOTYPE GENERATION
Full Name Short Name Reference
C-L. Chang. Finding Prototypes For Nearest Neighbor Classifiers. IEEE Transacti T. Kohonen. The Self-Organizative Map. Proceedings of the IEEE 78:9 (1990) 1464-1480. T. Kohonen. The Self-Organizative Map. Proceedings of the IEEE 78:9 (1990) 1464-1480. T. Kohonen. The Self-Organizative Map. Proceedings of the IEEE 78:9 (1990) 1464-1480. T. Kohonen. The Self-Organizative Map. Proceedings of the IEEE 78:9 (1990) 1464-1480. J. Koplowitz, T.A. Brown. On the relation of performance to editing in nearest neighbor rules. Pattern Recognition 13 (1981) 251-255. S. Geva, J. Site. Adaptive nearest neighbor pattern classifier. IEEE Transactions on Neural Networks 2:2 (1991) 318-322. Q. Xie, C.A. Laszlo. Vector quantization technique for nonparametric classifier design. IEEE Transactions on Pattern Analysis and Machine Intelligence 15:12 (1993) 1326-1330. C.H. Chen, A. Jzwik. A sample set condensation algorithm for the class sensitive artificial neural network. Pattern Recognition Letters 17 (1996) 819-823. Y. Hamamoto, S. Uchimura, S. Tomita. A bootstrap technique for nearest neighbor classifier design. IEEE Transactions on Pattern Analysis and Machine Intelligence 19:1 (1997) 73-79. C. Decaestecker. Finding prototypes for nearest neighbour classification by means of gradient descent and deterministic annealing. Pattern Recognition 30:2 (1997) 281-288. R. Odorico. Learning vector quantization with training count (LVQTC). Neural Networks 10:6 (1997) 1083-1088. J.C. Bezdek, T.R. Reichherzer, G.S. Lim, Y. Attikiouzel. Multiple prototype classifier design. IEEE Transactions on Systems, Man and Cybernetics C 28:1 (1998) 67-69. R.A. Mollineda, F.J. Ferri, E. Vidal. A merge-based condensing strategy for multiple prototype classifiers. IEEE Transactions on Systems, Man and Cybernetics B 32:5 (2002) 662-668. W. Lam, C.K. Keung, D. Liu. Discovering useful concept prototypes for classification based on filtering and abstraction. IEEE Transactions on Prototype Nearest PNN-C Neighbor Learning Vector Quantization 1 Learning Vector Quantization 2 Learning Vector Quantization 2.1 Learning Vector Quantization 3 Generalized Editing using Nearest Neighbor Decision Surface Mapping Vector Quantization LVQ1-C LVQ2-C LVQ2_1-C LVQ3-C GENN-C
DSM-C
VQ-C
Chen Algorithm
Chen-C
BTS3-C
MSE-C
LVQTC-C
MCA-C
GMCA-C
26
Pattern Analysis and Machine Intelligence 14:8 (2002) 1075-1090. Depuration Algorithm Depur-C J.S. Snchez, R. Barandela, A.I. Mrques, R. Alejo, J. Badenas. Analysis of new techniques to obtain quaylity training sets. Pattern Recognition Letters 24 (2003) 1015-1022. S.-W. Kim, A. Oomenn. A brief taxonomy and ranking of creative prototype reduction schemes. Pattern Analysis and Applications 6 (2003) 232244. J.S. Snchez. High training set size reduction by space partitioning and prototype abstraction. Pattern Recognition 37 (2004) 1561-1564.
HYB-C
High training set RSP-C size reduction by space partitioning and prototype abstraction Evolutionary ENPC-C Nearest Prototype Classifier Adaptive Vector Quantization AVQ-C
F. Fernndez, P. Isasi. Evolutionary design of nearest prototype classifiers. Journal of Heuristics 10:4 (2004) 431-454. C.-W. Yen, C.-N. Young, M.L. Nagurka. A vector quantization method for nearest neighbor classifier design. Pattern Recognition Letters 25 (2004) 725731. J. Li, M.T. Manry, C. Yu, D.R. Wilson. Prototype classifier design with pruning. International Journal on Artificial Intelligence Tools 14:1-2 (2005) 261280. T. Raicharoen, C. Lursinsap. A divide-and-conquer approach to the pairwise opposite class-nearest neighbor (poc-nn) algorithm. Pattern Recoginiton Letters 26 (2005) 1554-1567. M. Lozano, J.M. Sotoca, J.S. Snchez, F. Pla, E. Pekalska, R.P.W. Duin. Experimental study on prototype optimisation algorithms for prototypebased classification in vector spaces. Pattern Recognition 39:10 (2006) 1827-1838. H.A. Fayed, S.R. Hashem, A.F. Atiya. Selfgenerating prototypes for pattern classification. Pattern Recognition 40:5 (2007) 1498-1509. A. Cervantes, I. Galvn, P. Isasi. An Adaptive Michigan Approach PSO for Nearest Prototype Classification. 2nd International Work-Conference on the Interplay Between Natural and Artificial Computation (IWINAC07). LNCS 4528, Springer 2007, La Manga del Mar Menor (Spain, 2007) 287296. U. Garain. Prototype reduction using an artificial immune model. Pattern Analysis and Applications 11:3-4 (2008) 353-363. L. Nanni, A. Lumini. Particle swarm optimization for prototype reduction. Neurocomputing 72:4-6 (2009) 1092-1097.
LVQPRU-C
Pairwise Opposite POC-NN-C Class Nearest Neighbor Adaptive Condensing Algorithm Based on Mixtures of Gaussians Self-Generating Prototypes MixtGauss-C
SGP-C
PSCSA-C
PSO-C
27
NSC-C
C.J. Veenman, M.J.T. Reinders. The nearest subclass classifier: A compromise between the nearest mean and nearest neighbor classifier. IEEE Transactions on Pattern Analysis and Machine Intelligence 27:9 (2005) 1417-1429. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916. I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
Differential Evolution
DE-C
Scale Factor Local SFLSDE-C Search in Differential Evolution Self-Adaptive Differential Evolution Adaptive Differential Evolution with Optional External Archive Differential Evolution using a NeighborhoodBased Mutation Operator Hybrid Iterative Case Filtering + Learning Vector Quantization 3 Hybrid Iterative Case Filtering + Particle Swarm Optimization SADE-C
JADE-C
DEGL-C
ICFLVQ3-C
ICFPSO-C
Hybrid Iterative ICFSFLSDE-C Case Filtering + Scale Factor Local Search in Differential Evolution Hybrid SteadySSMALVQ3-C State Memetic Algorithm for Instance Selection + Learning Vector Quantization 3 Hybrid SteadySSMAPSO-C State Memetic Algorithm for Instance Selection + Particle Swarm Optimization
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
28
Hybrid SteadySSMASFLSDE-C State Memetic Algorithm for Instance Selection + Scale Factor Local Search in Differential Evolution Hybrid Decremental Reduction Optimization Procedure 3 + Learning Vector Quantization 3 Hybrid Decremental Reduction Optimization Procedure 3 + Particle Swarm Optimization DROP3LVQ3-C
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
DROP3PSO-C
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
Hybrid DROP3SFLSDE-C Decremental Reduction Optimization Procedure 3 + Scale Factor Local Search in Differential Evolution Iterative IPLDE-C Prototype Adjustment based on Differential Evolution
I. Triguero, S. Garca, F. Herrera. Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification. Pattern Recognition 44:4 (2011) 901-916.
I. Triguero, S. Garca, F. Herrera. IPADE: Iterative Prototype Adjustment for Nearest Neighbor Classification. IEEE Transactions on Neural Networks 21:12 (2010) 1984-1990.
LAZY LEARNING
Full Name
K-Nearest Neighbors Classifier Adaptive KNN Classifier
Short Name
KNN-C
Reference
T.M. Cover, P.E. Hart. Nearest Neighbor Pattern Classification. IEEE Transactions on Information Theory 13 (1967) 21-27. J. Wang, P. Neskovic, L.N. Cooper. Improving nearest neighbor rule with a simple adaptative distance measure.. Pattern Recognition Letters 28 (2007) 207-213. J.G. Cleary, L.E. Trigg. K*: An instance-based learner using an entropic distance measure. Proceedings of the 12th International Conference on Machine Learning. (1995) 108-114. J.H. Friedman, R. Kohavi, Y. Tun. Lazy decision trees. Proceedings of the Thirteenth National
KNNAdaptive-C
K * Classifier
KStar-C
29
Conference on Artificial Intellgence. (1996) 717724. Nearest Mean classifier NM-C T. Hastie, R. Tibshirani, J. Friedman. The elements of statistical learning: Data mining, inference, and prediction. Springer-Verlag, 2001. ISBN: 0-38795284-5. C. Zhou, Y. Chen. Improving nearest neighbor classification with cam weighted distance. Pattern Recognition 39 (2006) 635-645. Q. Gao, Z. Wang. Center-based nearest neighbor classifier. Pattern Recognition 40 (2007) 346-349. R. Nock, M. Sebban, D. Bernard. A simple locally adaptive nearest neighbor rule with application to pollution forecasting. International Journal of Pattern Recognition and Artificial Intelligence 17 (2003) 1369-1382. J. Li, G. Dong, K. Ramamohanarao, L. Wong. DeEPs: A New Instance-Based Lazy Discovery and Classification System. Machine Learning 54 (2004) 99-124. J. Li, G. Dong, K. Ramamohanarao, L. Wong. DeEPs: A New Instance-Based Lazy Discovery and Classification System. Machine Learning 54 (2004) 99-124.
Cam weighted distance Nearest Neighbor Classifier Center Nearest Neighbor Classifier K Symmetrical Nearest Neighbor Classifier
CamNN-C
CenterNN
KSNN-C
Decision making Deeps-C by Emerging Patterns Classifier Decision making DeepsNN-C by Emerging Patterns Classifier + Nearest Neighbor Classifier Lazy Bayesian Rules classifier Integrated Decremental Instance Based Learning LBR-C IDIBL
Z. Zheng, G.I. Webb. Lazy Learning of Bayesian Rules. Machine Learning 41 (2000) 53-87. D.R. Wilson, T.R. Martinez. An Integrated Instance-Based Learning Algorithm. Computational Intelligence 16:1 (2000) 1-28.
WEIGHTING METHODS
Full Name Short Name Reference
R. Paredes, E. Vidal. Learning weighted metrics to minimize nearest-neighbor classification error. IEEE Transactions on Pattern Analysis and Machine Intelligence 28:7 (2006) 1100-1110. R. Paredes, E. Vidal. Learning weighted metrics to minimize nearest-neighbor classification error. IEEE Transactions on Pattern Analysis and Machine Intelligence 28:7 (2006) 1100-1110. R. Paredes, E. Vidal. Learning weighted metrics to minimize nearest-neighbor classification error. IEEE Transactions on Pattern Analysis and Machine Intelligence 28:7 (2006) 1100-1110. Prototype PW-C weigthed classifier
CW-C
30
Short Name
RBFN-C
Reference
D.S. Broomhead, D. Lowe. Multivariable Functional Interpolation and Adaptive Networks. Complex Systems 11 (1988) 321-355.
Incr-RBFN-C
J. Plat. A Resource Allocating Network for Function Interpolation. Neural Computation 3:2 (1991) 213225.
SONN-C
I.G. Smotroff, D.H. Friedman, D. Connolly. Self Organizing Modular Neural Networks. Seattle International Joint Conference on Neural Networks (IJCNN'91). Seattle (USA, 1991) 187-192. F. Moller. A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks 6 (1990) 525-533. B. Widrow, M.A. Lehr. 30 Years of Adaptive Neural Networks: Perceptron, Madaline, and Backpropagation. Proceedings of the IEEE 78:9 (1990) 1415-1442.
MLP-CG-C
Decremental Radial Basis Function Neural Network for Classification Problems Ensemble Neural Network for Classification Problems Learning Vector Quantization for Classification Problems Evolutionary Radial Basis Function Neural Networks Improved Resilient backpropagation Plus
Decr-RBFN-C
D.S. Broomhead, D. Lowe. Multivariable Functional Interpolation and Adaptive Networks. Complex Systems 11 (1988) 321-355.
Ensemble-C
N. Garca-Pedrajas, C. Garca-Osorio, C. Fyfe. Nonlinear Boosting Projections for Ensemble Construction. Journal of Machine Learning Research 8 (2007) 1-33. J.C. Bezdek, L.I. Kuncheva. Nearest prototype classifier designs: An experimental study. International Journal of Intelligent Systems 16:12 (2001) 1445-1473. V.M. Rivas, J.J. Merelo, P.A. Castillo, M.G. Arenas, J.G. Castellano. Evolving RBF neural networks for time-series forecasting with EvRBF. Information Sciences 165:3-4 (2004) 207-220. C. Igel, M. Husken. Empirical evaluation of the improved Rprop learning algorithm. Neurocomputing 50 (2003) 105-123. L.R. Leerink, C.L. Giles, B.G. Horne, M.A. Jabri. Learning with Product Units. In: D. Touretzky, T. Leen (Eds.) Advances in Neural Information Processing Systems, 1995, 537-544.
LVQ-C
EvRBFN-C
iRProp+-C
31
MLP-BP-C
R. Rojas, J. Feldman. Neural Networks: A Systematic Introduction . Springer-Verlag, Berlin, New-York, 1996. ISBN: 978-3540605058.
Short Name
NNEP-C
Reference
F.J. Martnez-Estudillo, C. Hervs-Martnez, P.A. Gutirrez, A.C. Martnez-Estudillo. Evolutionary Product-Unit Neural Networks Classifiers. Neurocomputing 72:1-3 (2008) 548-561. G.F. Miller, P.M. Todd, S.U. Hedge. Designing Neural Networks Using Genetic Algorithms. 3rd International Conference on Genetic Algorithm and Their Applications. George Mason University (USA, 1989) 379-384. X. Yao. Evolving Artificial Neural Networks. Proceedings of the IEEE 87:9 (1999) 1423-1447.
Short Name
C_SVM-C NU_SVM-C
Reference
C. Cortes, V. Vapnik. Support vector networks. Machine Learning 20 (1995) 273-297. B. Scholkopf, A.J. Smola, R. Williamson, P.L. Bartlett. New support vector algorithms. Neural Computation 12 (2000) 1207-1245. J. Platt. Fast Training of Support Vector Machines using Sequential Minimal Optimization. In: B. Schoelkopf, C. Burges, A. Smola (Eds.) Advances in Kernel Methods - Support Vector Learning, 1998, 185-208. S.S. Keerthi, S.K. Shevade, C. Bhattacharyya, K.R.K. Murthy. Improvements to Platt's SMO Algorithm for SVM Classifier Design. Neural Computation 13:3 (2001) 637-649. T. Hastie, R. Tibshirani. Classification by Pairwise Coupling. In: M.I. Jordan, M.J. Kearns, S.A. Solla (Eds.) Advances in Neural Information Processing Systems, 1998, 451-471.
SMO-C
STATISTICAL CLASSIFIERS
Full Name
Nave-Bayes
Short Name
NB-C
Reference
P. Domingos, M. Pazzani. On the optimality of the simple Bayesian classifier under zero-one loss. Machine Learning 29 (1997) 103-137. M.E. Maron. Automatic Indexing: An Experimental
32
Inquiry. Journal of the ACM (JACM) 8:3 (1961) 404-417. Linear Discriminant Analysis LDA-C G.J. McLachlan. Discriminant Analysis and Statistical Pattern Recognition. John Wiley and Sons, 2004. R.A. Fisher. The Use of Multiple Measurements in Taxonomic Problems. Annals of Eugenics 7 (1936) 179-188. J.H. Friedman. Regularized Discriminant Analysis. Journal of the American Statistical Association 84 (1989) 165-175. Kernel Classifier Kernel-C G.J. McLachlan. Discriminant Analysis and Statistical Pattern Recognition. John Wiley and Sons, 2004. J.S. Rustagi. Optimization Techniques in Statistics. Academic Press, 1994. G.J. McLachlan. Discriminant Analysis and Statistical Pattern Recognition. John Wiley and Sons, 2004. R.A. Fisher. The Use of Multiple Measurements in Taxonomic Problems. Annals of Eugenics 7 (1936) 179-188. J.H. Friedman. Regularized Discriminant Analysis. Journal of the American Statistical Association 84 (1989) 165-175. Least Mean Square Quadratic classifier PolQuadraticLMS-C J.S. Rustagi. Optimization Techniques in Statistics. Academic Press, 1994. S. le Cessie, J.C. van Houwelingen. Ridge Estimators in Logistic Regression. Applied Statistics 41:1 (1992) 191-201. S.W. Lin, S.C. Chen. PSOLDA: A particle swarm optimization approach for enhancing classification accuracy rate of linear discriminant analysis. Applied Soft Computing 9 (2009) 1008-1015.
LinearLMS-C
QDA-C
Multinomial Logistic-C logistic regression model with a ridge estimator Particle Swarm Optimization Linear Discriminant Analysis PSOLDA-C
33
Regression Algorithms
Short Name
FRSBM-R
Reference
L. Snchez. A Random Sets-Based Method for Identifying Fuzzy Models. Fuzzy Sets and Systems 98:3 (1998) 343-354. L.X. Wang, J.M. Mendel. Generating Fuzzy Rules by Learning from Examples. IEEE Transactions on Systems, Man and Cybernetics 22:6 (1992) 14141427.
WM-R
Short Name
TSK-IRL-R
Reference
O. Cordn, F. Herrera. A Two-Stage Evolutionary Process for Designing TSK Fuzzy Rule-Based Systems. IEEE Transactions on Systems, Man and Cybernetics, Part B: Cybernetics 29:6 (1999) 703715. O. Cordn, F. Herrera. A Three-Stage Evolutionary Process for Learning Descriptive and Approximate Fuzzy Logic Controller Knowledge Bases from Examples. International Journal of Approximate Reasoning 17:4 (1997) 369-407. L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators with SA Search to Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-191. O. Cordn, F. Herrera. Hybridizing Genetic Algorithms with Sharing Scheme and Evolution Strategies for Designing Approximate Fuzzy RuleBased Systems. Fuzzy Sets and Systems 118:2 (2001) 235-255. R. Alcal, J. Alcala-Fdez, J. Casillas, O. Cordn, F. Herrera. Local Identification of Prototypes for Genetic Learning of Accurate TSK Fuzzy RuleBased Systems.. International Journal of Intelligent Systems 22:9 (2007) 909-941. L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators with SA Search to Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-191.
Iterative Rule MOGUL-IRLSC-R Learning of Mamdani Rules Small Constrained Approach Fuzzy Learning based on Genetic Programming Grammar Operators Iterative Rule Learning of Mamdani Rules High Constrained Approach Learning TSKFuzzy Models Based on MOGUL GFS-GPG-R
MOGUL-IRLHC-R
MOGUL-TSK-R
Fuzzy Learning based on Genetic Programming Grammar Operators and Simulated Annealing Genetic Fuzzy Rule Learning, Thrift Algorithm
GFS-SP-R
Thrift-R
P. Thrift. Fuzzy logic synthesis with genetic algorithms. Proceedings of the Fourth International Conference on Genetic Algorithms (ICGA91). San Diego (United States of America, 1991) 509-513.
34
GFS-RB-MF-R
A. Homaifar, E. McCormick. Simultaneous Design of Membership Functions and Rule Sets for Fuzzy Controllers Using Genetic Algorithms. IEEE Transactions on Fuzzy Systems 3:2 (1995) 129139. O. Cordn, F. Herrera. A Three-Stage Evolutionary Process for Learning Descriptive and Approximate Fuzzy Logic Controller Knowledge Bases from Examples. International Journal of Approximate Reasoning 17:4 (1997) 369-407.
Iterative Rule Learning of Descriptive Mamdani Rules based on MOGUL SymbioticEvolution-based Fuzzy Controller design method Pittsburgh Fuzzy Classifier System #1
MOGUL-IRL-R
O. Cordn, F. Herrera. A Three-Stage Evolutionary Process for Learning Descriptive and Approximate Fuzzy Logic Controller Knowledge Bases from Examples. International Journal of Approximate Reasoning 17:4 (1997) 369-407. C.F. Juang, J.Y. Lin, C.-T. Lin. Genetic reinforcement learning through symbiotic evolution for fuzzy controller design. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 30:2 (2000) 290-302. B. Carse, T.C. Fogarty, A. Munro. Evolving fuzzy rule based controllers using genetic algorithms. Fuzzy Sets and Systems 80:3 (1996) 273-293.
SEFC-R
P_FCS1-R
Short Name
M5-R
Reference
J.R. Quinlan. Learning with Continuous Classes. 5th Australian Joint Conference on Artificial Intelligence (AI92). (Singapore, 1992) 343-348. I. Wang, I.H. Witten. Induction of model trees for predicting continuous classes. 9th European Conference on Machine Learning. Prague (Czech Republic, 1997) 128-137.
CART-R
L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone. Classification and Regression Trees. Chapman and Hall (Wadsworth, Inc.), 1984. J.R. Quinlan. Learning with Continuous Classes. Proceedings of the 5th Australian Joint Conference on Artificial Intelligence. (1992) 343-348. I. Wang, I.H. Witten. Induction of model trees for predicting continuous classes. Poster papers of the 9th European Conference on Machine Learning. Prague (Czech Republic, 1997) 128-137. G. Holmes, M. Hall, E. Frank. Generating Rule Sets from Model Trees. Proceedings of the 12th Australian Joint Conference on Artificial Intelligence: Advanced Topics in Artificial Intelligence. Springer-Verlag. Sydney (Australia, 1999) 1-12.
M5Rules-R
35
Short Name
GFS-Ling-T
Reference
O. Cordn, F. Herrera. A Three-Stage Evolutionary Process for Learning Descriptive and Approximate Fuzzy Logic Controller Knowledge Bases from Examples. International Journal of Approximate Reasoning 17:4 (1997) 369-407. F. Herrera, M. Lozano, J.L. Verdegay. Tuning Fuzzy Logic Controllers by Genetic Algorithms. International Journal of Approximate Reasoning 12 (1995) 299-315. O. Cordn, F. Herrera. A Three-Stage Evolutionary Process for Learning Descriptive and Approximate Fuzzy Logic Controller Knowledge Bases from Examples. International Journal of Approximate Reasoning 17:4 (1997) 369-407. H. Ishibuchi, K. Nozaki, N. Yamamoto, H. Tanaka. Selecting Fuzzy If-Then Rules for Classification Problems Using Genetic Algorithms. IEEE Transactions on Fuzzy Systems 3:3 (1995) 260270.
Approximative GFS-Aprox-T Genetic Tuning of FRBSs Genetic Selection of Linguistic Rule Bases GFS-RS-T
R. Alcal, O. Cordn, F. Herrera. Combining Rule Weight Learning and Rule Selection to Obtain Simpler and More Accurate Linguistic Fuzzy Models. In: J. Lawry, J.G. Shanahan, A.L. Ralescu (Eds.) Modelling with Words, LNCS 2873, 2003, 44-63. R. Alcal, O. Cordn, F. Herrera. Combining Rule Weight Learning and Rule Selection to Obtain Simpler and More Accurate Linguistic Fuzzy Models. In: J. Lawry, J.G. Shanahan, A.L. Ralescu (Eds.) Modelling with Words, LNCS 2873, 2003, 44-63. D. Park, A. Kandel. Genetic-Based New Fuzzy Reasoning Model with Application to Fuzzy Control. IEEE Transactions on System, Man and Cybernetics, Part B: Cybernetics 24:1 (1994) 3947. R. Alcal, J. Alcala-Fdez, M.J. Gacto, F. Herrera. Rule Base Reduction and Genetic Tuning of Fuzzy Systems based on the Linguistic 3-Tuples Representation. Soft Computing 11:5 (2007) 401419. R. Alcal, J. Alcala-Fdez, F. Herrera. A Proposal for the Genetic Lateral Tuning of Linguistic Fuzzy Systems and Its Interaction With Rule Selection. IEEE Transactions on Fuzzy Systems 15:4 (2007) 616-635. R. Alcal, J. Alcala-Fdez, M.J. Gacto, F. Herrera. Rule Base Reduction and Genetic Tuning of Fuzzy Systems based on the Linguistic 3-Tuples Representation. Soft Computing 11:5 (2007) 401-
GFS-Weight-RS-T
GFS-GB-NFRM-T
Local Genetic Lateral and Amplitude Tuning of FRBSs Local Genetic Lateral Tuning of FRBSs
GFS-LLA-T
GFS-LL-T
36
of FRBSs Local Genetic GFS-LLRS-T Lateral Tuning with rule selection of FRBSs Global Genetic GFS-GLA-T Lateral and Amplitude-Tuning of FRBSs Global Genetic Lateral Tuning of FRBSs GFS-GL-T
419. R. Alcal, J. Alcala-Fdez, F. Herrera. A Proposal for the Genetic Lateral Tuning of Linguistic Fuzzy Systems and Its Interaction With Rule Selection. IEEE Transactions on Fuzzy Systems 15:4 (2007) 616-635. R. Alcal, J. Alcala-Fdez, M.J. Gacto, F. Herrera. Rule Base Reduction and Genetic Tuning of Fuzzy Systems based on the Linguistic 3-Tuples Representation. Soft Computing 11:5 (2007) 401419. R. Alcal, J. Alcala-Fdez, F. Herrera. A Proposal for the Genetic Lateral Tuning of Linguistic Fuzzy Systems and Its Interaction With Rule Selection. IEEE Transactions on Fuzzy Systems 15:4 (2007) 616-635. R. Alcal, J. Alcala-Fdez, M.J. Gacto, F. Herrera. Rule Base Reduction and Genetic Tuning of Fuzzy Systems based on the Linguistic 3-Tuples Representation. Soft Computing 11:5 (2007) 401419. R. Alcal, J. Alcala-Fdez, F. Herrera. A Proposal for the Genetic Lateral Tuning of Linguistic Fuzzy Systems and Its Interaction With Rule Selection. IEEE Transactions on Fuzzy Systems 15:4 (2007) 616-635.
Global Genetic GFS-GLARS-T Lateral and Amplitude-Tuning with rule selection of FRBSs Global Genetic GFS-GLRS-TS Lateral Tuning with rule selection of FRBSs
Short Name
MLP-CG-R
Reference
F. Moller. A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks 6 (1990) 525-533.
RBFN-R
D.S. Broomhead, D. Lowe. Multivariable Functional Interpolation and Adaptive Networks. Complex Systems 11 (1998) 321-355. J. Plat. A Resource Allocating Network for Function Interpolation. Neural Computation 3:2 (1991) 213225. I.G. Smotroff, D.H. Friedman, D. Connolly. Self Organizing Modular Neural Networks. Seattle International Joint Conference on Neural Networks (IJCNN'91). Seattle (USA, 1991) 187-192. D.S. Broomhead, D. Lowe. Multivariable Functional Interpolation and Adaptive Networks. Complex Systems 11 (1988) 321-355.
Incr-RBFN-R
SONN-R
Decr-RBFN-R
37
Multilayer Perceptron with Backpropagation Based Training Improved Resilient backpropagation Plus
MLP-BP-R
R. Rojas, J. Feldman. Neural Networks: A Systematic Introduction . Springer-Verlag, Berlin, New-York, 1996. ISBN: 978-3540605058. C. Igel, M. Husken. Empirical evaluation of the improved Rprop learning algorithm. Neurocomputing 50 (2003) 105-123. J.H. Wang, Y.W. Yu, J.H. Tsai. On the internal representations of product units. Neural Processing Letters 12:3 (2000) 247-254.
iRProp+-R
Ensemble-R
N. Garca-Pedrajas, C. Garca-Osorio, C. Fyfe. Nonlinear Boosting Projections for Ensemble Construction. Journal of Machine Learning Research 8 (2007) 1-33.
Short Name
EPSILON_SVR-R
Reference
R.E. Fan, P.H. Chen, C.J. Lin. Working set selection using the second order information for training SVM. Journal of Machine Learning Research 6 (2005) 1889-1918. R.E. Fan, P.H. Chen, C.J. Lin. Working set selection using the second order information for training SVM. Journal of Machine Learning Research 6 (2005) 1889-1918.
NU-SVR
NU_SVR-R
38
Symbolic Fuzzy GFS-GSP-R Learning based on Genetic Programming Grammar Operators and Simulated Annealing Symbolic Fuzzy GFS-GP-R Learning based on Genetic Programming Symbolic FuzzyGFS-SAP-Sym-R Valued Data Learning based on Genetic Programming Grammar Operators and Simulated Annealing
L. Snchez, I. Couso. Fuzzy Random VariablesBased Modeling with GA-P Algorithms. In: B. Bouchon, R.R. Yager, L. Zadeh (Eds.) Information, Uncertainty and Fusion, 2000, 245-256. L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators with SA Search to Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-191. L. Snchez, I. Couso. Fuzzy Random VariablesBased Modeling with GA-P Algorithms. In: B. Bouchon, R.R. Yager, L. Zadeh (Eds.) Information, Uncertainty and Fusion, 2000, 245-256. L. Snchez, I. Couso. Fuzzy Random VariablesBased Modeling with GA-P Algorithms. In: B. Bouchon, R.R. Yager, L. Zadeh (Eds.) Information, Uncertainty and Fusion, 2000, 245-256. L. Snchez, I. Couso, J.A. Corrales. Combining GP Operators with SA Search to Evolve Fuzzy Rule Based Classifiers. Information Sciences 136:1-4 (2001) 175-191.
STATISTICAL REGRESSION
Full Name
Least Mean Squares Linear Regression
Short Name
LinearLMS-R
Reference
J.S. Rustagi. Optimization Techniques in Statistics. Academic Press, 1994. J.S. Rustagi. Optimization Techniques in Statistics. Academic Press, 1994.
39
Imbalanced Classification
OVER-SAMPLING METHODS
Full Name Short Name Reference
N.V. Chawla, K.W. Bowyer, L.O. Hall, W.P. Kegelmeyer. SMOTE: Synthetic Minority Oversampling TEchnique. Journal of Artificial Intelligence Research 16 (2002) 321-357. G.E.A.P.A. Batista, R.C. Prati, M.C. Monard. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations 6:1 (2004) 20-29. G.E.A.P.A. Batista, R.C. Prati, M.C. Monard. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations 6:1 (2004) 20-29. Synthetic Minority SMOTE-I Over-sampling TEchnique Synthetic Minority SMOTE_ENN-I Over-sampling TEchnique + Edited Nearest Neighbor Synthetic Minority SMOTE_TL-I Over-sampling TEchnique + Tomek's modification of Condensed Nearest Neighbor ADAptive SYNthetic Sampling ADASYN-I
H. He, Y. Bai, E.A. Garcia, S. Li. ADASYN: Adaptive Synthetic Sampling Approach for Imbalanced Learning. 2008 International Joint Conference on Neural Networks (IJCNN08). Hong Kong (Hong Kong Special Administrative Region of the Peo, 2008) 1322-1328. H. Han, W.Y. Wang, B.H. Mao. Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. 2005 International Conference on Intelligent Computing (ICIC05). LNCS 3644, Springer 2005, Hefei (China, 2005) 878-887. C. Bunkhumpornpat, K. Sinapiromsaran, C. Lursinsap. Safe-level-SMOTE: Safe-level-synthetic minority over-sampling technique for handling the class imbalanced problem. Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD09). LNCS 5476, Springer 2009, Bangkok (Thailand, 2009) 475-482. G.E.A.P.A. Batista, R.C. Prati, M.C. Monard. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations 6:1 (2004) 20-29. S. Tang, S. Chen. The Generation Mechanism of Synthetic Minority Class Examples. 5th Int. Conference on Information Technology and Applications in Biomedicine (ITAB 2008). Shenzhen (China, 2008) 444-447. J. Stefanowski, S. Wilk. Selective pre-processing of imbalanced data for improving classification performance. 10th International Conference in Data Warehousing and Knowledge Discovery (DaWaK2008). LNCS 5182, Springer 2008, Turin
BorderlineBorderline_SMOTE-I Synthetic Minority Over-sampling TEchnique Safe Level Safe_Level_SMOTE-I Synthetic Minority Over-sampling TEchnique
Random oversampling
ROS-I
Adjusting the ADOMS-I Direction Of the synthetic Minority clasS examples Selective Preprocessing of Imbalanced Data SPIDER-I
40
(Italy, 2008) 283-292. Aglomerative Hierarchical Clustering Selective Preprocessing of Imbalanced Data 2 Hybrid Preprocessing using SMOTE and Rough Sets Theory AHC-I G. Cohen, M. Hilario, H. Sax, S. Hugonnet, A. Geissbuhler. Learning from imbalanced data in surveillance of nosocomial infection. Artificial Intelligence in Medicine 37 (2006) 7-18. K. Napierala, J. Stefanowski, S. Wilk. Learning from Imbalanced Data in Presence of Noisy and Borderline Examples. 7th International Conference on Rough Sets and Current Trends in Computing (RSCTC2010). Warsaw (Poland, 2010) 158-167. E. Ramentol, Y. Caballero, R. Bello, F. Herrera. SMOTE-RSB*: A Hybrid Preprocessing Approach based on Oversampling and Undersampling for High Imbalanced Data-Sets using SMOTE and Rough Sets Theory. Knowledge and Information Systems (2011) In press.
SPIDER2-I
SMOTE_RSB-I
UNDER-SAMPLING METHODS
Full Name
Tomek's modification of Condensed Nearest Neighbor Condensed Nearest Neighbor Random undersampling
Short Name
TL-I
Reference
I. Tomek. Two modifications of CNN. IEEE Transactions on Systems, Man and Cybernetics 6 (1976) 769-772. P.E. Hart. The Condensed Nearest Neighbour Rule. IEEE Transactions on Information Theory 14:5 (1968) 515-516. G.E.A.P.A. Batista, R.C. Prati, M.C. Monard. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations 6:1 (2004) 20-29. M. Kubat, S. Matwin. Addressing the curse of imbalanced training sets: one-sided selection. 14th International Conference on Machine Learning (ICML97). Tennessee (USA, 1997) 179-186. G.E.A.P.A. Batista, R.C. Prati, M.C. Monard. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations 6:1 (2004) 20-29.
CNN-I
RUS-I
OSS-I
Condensed Nearest Neighbor + Tomek's modification of Condensed Nearest Neighbor Neighborhood Cleaning Rule
CNNTL-I
NCL-I
J. Laurikkala. Improving Identification of Difficult Small Classes by Balancing Class Distribution . 8th Conference on AI in Medicine in Europe (AIME01). LNCS 2001, Springer 2001, Cascais (Portugal, 2001) 63-66. S. Yen, Y. Lee. Under-sampling approaches for improving prediction of the minority class in an imbalanced dataset. International Conference on Intelligent Computing (ICIC06). Kunming (China, 2006) 731-740.
SBC-I
41
CPM-I
K. Yoon, S. Kwek. An unsupervised learning approach to resolving the data imbalanced issue in supervised learning problems in functional genomics. 5th International Conference on Hybrid Intelligent Systems (HIS05). Rio de Janeiro (Brazil, 2005) 303-308.
COST-SENSITIVE CLASSIFICATION
Full Name
C4.5 CostSensitive
Short Name
C45CS-I
Reference
K.M. Ting. An instance-weighting method to induce cost-sensitive trees. IEEE Transactions on Knowledge and Data Engineering 14:3 (2002) 659665. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
NNCS-I
Z.-H. Zhou, X.-Y. Liu. Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Transactions on Knowledge and Data Engineering 18:1 (2006) 6377. R. Rojas, J. Feldman. Neural Networks: A Systematic Introduction . Springer-Verlag, Berlin, New-York, 1996. ISBN: 978-3540605058.
C-SVM CostSensitive
C_SVMCS-I
K. Veropoulos, N. Cristianini, C. Campbell. Controlling the sensitivity of support vector machines. 16th International Joint Conferences on Artificial Intelligence (IJCAI99). Stockholm (Sweden, 1999) 281-288. Y. Tang, Y.-Q. Zhang, N.V. Chawla. SVMs modeling for highly imbalanced classification. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 39:1 (2009) 0-288.
42
Y. Freund, R.E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences 55:1 (1997) 119-139. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Adaptive Boosting AdaBoostM2-I Second MultiClass Extension with C4.5 Decision Tree as Base Classifier
R.E. Schapire, Y. Singer. Improved boosting algorithms using confidence-rated predictions. Machine Learning 37 (1999) 297-336. Y. Freund, R.E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences 55:1 (1997) 119-139. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Bootstrap Bagging-I Aggregating with C4.5 Decision Tree as Base Classifier BalanceCascade BalanceCascade-I Ensemble with C4.5 Decision Tree as Base Classifier
L. Breiman. Bagging predictors. Machine Learning 24 (1996) 123-140. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. X.-Y. Liu, J. Wu, Z.-H. Zhou. Exploratory undersampling for class-imbalance learning. IEEE Transactions on Systems, Man, and Cybernetics, Part B 39:2 (2009) 539-550. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Boosting with Data Generation for Imbalanced Data with C4.5 Decision Tree as Base Classifier
DataBoost-IM-I
H. Guo, H.L. Viktor. Learning from imbalanced data sets with boosting and data generation: the DataBoost-IM approach. SIGKDD Explorations 6:1 (2004) 30-39. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. X.-Y. Liu, J. Wu, Z.-H. Zhou. Exploratory undersampling for class-imbalance learning. IEEE Transactions on Systems, Man, and Cybernetics, Part B 39:2 (2009) 539-550. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Integrating IIVotes-I Selective Preprocessing of Imbalanced Data with Ivotes Ensemble with C4.5 Decision Tree as Base Classifier
J. Blaszczynski, M. Deckert, J. Stefanowski, S. Wilk. Integrating selective pre-processing of imbalanced data with ivotes ensemble. 7th International Conference on Rough Sets and Current Trends in Computing (RSCTC2010). LNCS 6086, Springer 2010, Warsaw (Poland, 2010) 148157. L. Breiman. Pasting small votes for classification in large databases and on-line. Machine Learning 36 (1999) 85-103. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
S. Wang, X. Yao. Diversity analysis on imbalanced data sets by using ensemble models. IEEE
43
Symposium Series on Computational Intelligence and Data Mining (IEEE CIDM 2009). Nashville TN (USA, 2009) 324-331. M. Galar, A. Fernndez, E. Barrenechea, H. Bustince, F. Herrera. A Review on Ensembles for the Class Imbalance Problem: Bagging-, Boosting, and Hybrid-Based Approaches. IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews (2011) In press. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Modified Synthetic MSMOTEBoost-I Minority Oversampling TEchnique Boost with C4.5 Decision Tree as Base Classifier Over-sampling OverBagging-I Minority Classes Bagging with C4.5 Decision Tree as Base Classifier
S. Hu, Y. Liang, L. Ma, Y. He. MSMOTE: Improving classification performance when training data is imbalanced. 2nd International Workshop on Computer Science and Engineering (WCSE 2009). Qingdao (China, 2009) 13-17. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. S. Wang, X. Yao. Diversity analysis on imbalanced data sets by using ensemble models. IEEE Symposium Series on Computational Intelligence and Data Mining (IEEE CIDM 2009). Nashville TN (USA, 2009) 324-331. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Over-sampling OverBagging2-I Minority Classes Bagging 2 with C4.5 Decision Tree as Base Classifier
S. Wang, X. Yao. Diversity analysis on imbalanced data sets by using ensemble models. IEEE Symposium Series on Computational Intelligence and Data Mining (IEEE CIDM 2009). Nashville TN (USA, 2009) 324-331. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Random UnderRUSBoost-I Sampling Boosting with C4.5 Decision Tree as Base Classifier
C. Seiffert, T. Khoshgoftaar, J. Van Hulse, A. Napolitano. Rusboost: A hybrid approach to alleviating class imbalance. IEEE Transactions on Systems, Man and Cybernetics, Part A 40:1 (2010) 185-197. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
Synthetic Minority SMOTEBagging-I Over-sampling TEchnique Bagging with C4.5 Decision Tree as Base Classifier Synthetic Minority SMOTEBoost-I Over-sampling TEchnique Boosting with C4.5 Decision Tree as Base Classifier
S. Wang, X. Yao. Diversity analysis on imbalanced data sets by using ensemble models. IEEE Symposium Series on Computational Intelligence and Data Mining (IEEE CIDM 2009). Nashville TN (USA, 2009) 324-331. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. N.V. Chawla, A. Lazarevic, L.O. Hall, K.W. Bowyer. SMOTEBoost: Improving prediction of the minority class in boosting. 7th European Conference on Principles and Practice of Knowledge Discovery in Databases (PKDD 2003). Cavtat Dubrovnik (Croatia, 2003) 107-119.
44
J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. Under-sampling UnderBagging-I Minority Classes Bagging with C4.5 Decision Tree as Base Classifier Under-sampling UnderBagging2-I Minority Classes Bagging 2 with C4.5 Decision Tree as Base Classifier Under-sampling UnderOverBagging-I Minority Classes Bagging to Oversampling Minority Classes Bagging with C4.5 Decision Tree as Base Classifier R. Barandela, R.M. Valdovinos, J.S. Snchez. New applications of ensembles of classifiers. Pattern Analysis and Applications 6 (2003) 245-256. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. R. Barandela, R.M. Valdovinos, J.S. Snchez. New applications of ensembles of classifiers. Pattern Analysis and Applications 6 (2003) 245-256. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993. S. Wang, X. Yao. Diversity analysis on imbalanced data sets by using ensemble models. IEEE Symposium Series on Computational Intelligence and Data Mining (IEEE CIDM 2009). Nashville TN (USA, 2009) 324-331. J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kauffman, 1993.
45
Subgroup Discovery
SUBGROUP DISCOVERY
Full Name Short Name Reference
N. Lavrac, B. Kavsek, P. Flach, L. Todorovski.. Subgroup Discovery with CN2-SD. Journal of Machine Learning Research 5 (2004) 153-188. B. Kavsek, N. Lavrac. APRIORI-SD: Adapting Association Rule Learning to Subgroup Discovery. Applied Artificial Intelligence 20:7 (2006) 543583. D. Gambergr, N. Lavrac. Expert-Guided Subgroup Discovery: Methodology and Application. Journal of Artificial Intelligence Research 17 (2002) 501-527. M.J. del Jesus, P. Gonzlez, F. Herrera, M. Mesonero. Evolutionary Fuzzy Rule Induction Process for Subgroup Discovery: A case study in marketing. IEEE Transactions on Fuzzy Systems 15:4 (2007) 578-592. C.J. Carmona, P. Gonzlez, M.J. del Jesus, F. Herrera. Non-dominated Multi-objective Evolutionary algorithm based on Fuzzy rules extraction for Subgroup Discovery. 4th International Conference on Hybrid Artificial Intelligence Systems (HAIS09). LNCS 5572, Springer 2009, Salamanca (Spain, 2009) 573-580. F.J. Berlanga, M.J. del Jesus, P. Gonzlez, F. Herrera, M. Mesonero. Multiobjective Evolutionary Induction of Subgroup Discovery Fuzzy Rules: A Case Study in Marketing. 6th Industrial Conference on Data Mining. LNCS 4065, Springer 2006, Leipzig (Germany, 2006) 337-349. M.J. del Jesus, P. Gonzlez, F. Herrera. Multiobjective Genetic Algorithm for Extracting Subgroup Discovery Fuzzy Rules. IEEE Symposium on Computational Intelligence in Multicriteria Decision Making. (2007) 0-57. SD-Map algorithm SDMap-SD M. Atzmueller, F. Puppe. SD-Map - A Fast Algorithm for Exhaustive Subgroup Discovery. 17th European Conference on Machine Learning and 10th European Conference on Principles and Practice of Knowledge Discovery in Databases (ECML/PKDD 2006). LNCS 4213, Springer 2006, Berlin (Germany, 2006) 6-17. CN2 Algorithm for CN2-SD Subgroup Discovery Apriori Algorithm for Subgroup Discovery Subgroup Discovery Algorithm Subgroup Discovery Iterative Genetic Algorithms Apriori-SD
SD-Algorithm-SD
SDIGA-SD
Non-dominated NMEEF-SD Multi-Objective Evolutionary algorithm for Extracting Fuzzy rules in Subgroup Discovery MESDIF for Subgroup Discovery MESDIF-SD
46
Expectation Maximization Diverse Density K-Nearest Neighbors for Multiple Instance Learning
EMDD-M
KNN-MI-M
Grammar-Guided G3P-MI-M Genetic Programming for Multiple Instance Learning Axis Parallel Rectangle using Iterated Discrimination Axis Parallel Rectangle using positive vectors covering eliminating negative instances Axis Parallel Rectangle eliminating negative instances Axis Parallel Rectangle eliminating negative instances based on a kernel density estimate APR_Iterated Discrimination-M
APR_GFS_AllPositive-M
APR_GFS_ElimCount-M
T.G. Dietterich, R.H. Lathrop, T. Lozano-Prez. Solving the multiple instance problem with axisparallel rectangles. Artificial Intelligence 89 (1997) 31-71. T.G. Dietterich, R.H. Lathrop, T. Lozano-Prez. Solving the multiple instance problem with axisparallel rectangles. Artificial Intelligence 89 (1997) 31-71.
APR_GFS_Kde-M
47
Clustering Algorithms
CLUSTERING ALGORITHMS
Full Name
ClusterKMeans
Short Name
KMeans-CL
Reference
J.B. MacQueen. Some Methods for Classification and Analysis of Multivariate Observations. 5th Berkeley Symposium on Mathematical Statistics and Probability. Berkeley (USA, 1967) 281-297.
48
Association Rules
ASSOCIATION RULES
Full Name
Apriori
Short Name
Apriori-A
Reference
R. Srikant, R. Agrawal. Mining quantitative association rules in large relational tables. ACM SIGMOD International Conference on Management of Data. Montreal Quebec (Canada, 1996) 1-12. C. Borgelt. Efficient implementations of Apriori and Eclat. Workshop of Frequent Item Set Mining Implementations (FIMI 2003). Florida (USA, 2003) 280-296.
Alatasetal-A
B. Alatas, E. Akin. An efficient genetic algorithm for automated mining of both positive and negative quantitative association rules. Soft Computing 10 (2006) 230-237.
Evolutionary EARMGA-A Association Rules Mining with Genetic Algorithm Equivalence CLAss Eclat-A Transformation
X. Yan, Ch. Zhang, S. Zhang. Genetic algorithmbased strategy for identifying association rules without specifying actual minimum support. Expert Systems with Applications 36:2 (2009) 3066-3076. M.J. Zaki. Scalable Algorithms for Association Mining. IEEE Transactions on Knowledge and Data Engineering 12:3 (2000) 372-390. C. Borgelt. Efficient implementations of Apriori and Eclat. Workshop of Frequent Item Set Mining Implementations (FIMI 2003). Florida (USA, 2003) 280-296.
FPgrowth-A
J. Han, J. Pei, Y. Yin, R. Mao. Mining frequent patterns without candidate generation: A frequentpattern tree approach. Data Mining and Knowledge Discovery 8:1 (2004) 53-87. J. Mata, J.L. Alvarez, J.C. Riquelme. Discovering numeric association rules via evolutionary algorithm. Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD). Springer, Heidelberg. Hong Kong (China, 2001) 40-51. J. Mata, J.L. Alvarez, J.C. Riquelme. An evolutionary algorithm to discover numeric association rules. ACM Symposium on Applied Computing. Madrid (Spain, 2002) 0-594.
GAR-A
GENAR-A
J. Mata, J.L. Alvarez, J.C. Riquelme. Mining numeric association rules with genetic algorithms. 5th International Conference on Artificial Neural Networks and Genetic Algorithms (ICANNGA). Taipei (Taiwan, 2001) 264-267. J. Alcala-Fdez, R. Alcal, M.J. Gacto, F. Herrera. Learning the membership function contexts for mining fuzzy association rules by using genetic algorithms. Fuzzy Sets and Systems 160 (2009)
Alcala et al Method
Alcalaetal-A
49
905-921. Fuzzy Apriori FuzzyApriori-A T.-P. Hong, C.-S. Kuo, S.-C. Chi. Trade-off between computation time and number of rules for fuzzy mining from quantitative data. International Journal of Uncertainty, Fuzziness and KnowledgeBased Systems 9:5 (2001) 587-604. T.-P. Hong, C.-H. Chen, Y.-L. Wu, Y.-C. Lee. A GAbased fuzzy mining approach to achieve a trade-off between number of rules and suitability of membership functions. Soft Computing 10:11 (2006) 1091-1101.
GeneticFuzzyApriori-A
Genetic-Fuzzy GeneticFuzzyAprioriDC- T.-P. Hong, C.-H. Chen, Y.-C. Lee, Y.-L. Wu. Data Mining With A Genetic-Fuzzy Data Mining With Divide-andDivide-andConquer Strategy. IEEE Transactions on Conquer Strategy Evolutionary Computation 12:2 (2008) 252-265.
50
Statistical Tests
TEST ANALYSIS
Full Name
5x2 Cross validation F-test
Short Name
5x2CV-ST
Reference
T.G. Dietterich. Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms. Neural Computation 10:7 (1998) 1895-1923. F. Wilcoxon. Individual Comparisons by Ranking Methods. Biometrics 1 (1945) 80-83. J.P. Royston. Algorithm AS 181. Applied Statistics 31:2 (1982) 176-180.
Wilcoxon signed ranks test (for a single data-set) T-test Snedecor F-test Normality Shapiro-Wilk test Mann-Whitney Utest
Single-Wilcoxon-ST
D.R. Cox, D.V. Hinkley. Theoretical Statistics. Chapman and Hall, 1974. G.W. Snedecor, W.G. Cochran. Statistical Methods. Iowa State University Press, 1989. S.S. Shapiro, M.B. Wilk. An Analysis of Variance Test for Normality (complete samples). Biometrika 52:3-4 (1965) 591-611. H.B. Mann, D.R. Whitney. On a Test of Whether One of Two Random Variables is Stochastically Larger Than The Other. Annals of Mathematical Statistics 18 (1947) 50-60. F. Wilcoxon. Individual Comparisons by Ranking Methods. Biometrics 1 (1945) 80-83. J.P. Royston. Algorithm AS 181. Applied Statistics 31:2 (1982) 176-180.
MannWhitneyU-ST
Wilcoxon-ST
D. Sheskin. Handbook of parametric and nonparametric statistical procedures. Chapman and Hall/CRC, 2003. M. Friedman. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. Journal of the American Statistical Association 32:200 (1937) 675-701.
Quade-ST
D. Quade. Using weighted rankings in the analysis of complete blocks with additive block effects. Journal of the American Statistical Association 74 (1979) 680-683. W.J. Conover. Practical Nonparametric Statistics. Wiley, 1998.
J.L. Hodges, E.L. Lehmann. Ranks methods for combination of independent experiments in analysis of variance. Annals of Mathematical Statistics 33 (1962) 482-497. W.W. Daniel. Applied Nonparametric Statistics. Houghton Mifflin Harcourt, 1990.
R.G.D. Steel. A multiple comparison sign test: treatments versus control. Journal of American Statistical Association 54 (1959) 767-775.
51
Post-Hoc Procedures
D. Sheskin. Handbook of parametric and nonparametric statistical procedures. Chapman and Hall/CRC, 2003. M. Friedman. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. Journal of the American Statistical Association 32:200 (1937) 675-701.
Contrast estimation
Contrast-Test-ST
K. Doksum. Robust procedures for some linear models with one observation per cell. Annals of Mathematical Statistics 38 (1967) 878-883.
52
Shaffer Post Hoc Multiple-Test-ST procedure for NxN Statistical Tests Bergman Post Hoc Multiple-Test-ST procedure for NxN Statistical Tests
J.P. Shaffer. Modified sequentially rejective multiple test procedures. Journal of the American Statistical Association 81:395 (1986) 826-831. G. Bergmann, G. Hommel. Improvements of general multiple test procedures for redundant systems of hypotheses. In: P. Bauer, G. Hommel, E. Sonnemann (Eds.) Multiple Hypotheses Testing, 1988, 100-115.
53