Weka & Rapid Miner Tutorial: by Chibuike Muoh
Weka & Rapid Miner Tutorial: by Chibuike Muoh
Weka & Rapid Miner Tutorial: by Chibuike Muoh
By Chibuike Muoh
WEKA:: Introduction
WEKA:: Installation
If you are interested in modifying/extending weka there is a developer version that includes the source code
setenv WEKAHOME /usr/local/weka/weka-3-0-2 setenv CLASSPATH $WEKAHOME/weka.jar:$CLASSPATH
Routines are implemented as classes and logically arranged in packages Comes with an extensive GUI interface
Weka routines can be used stand alone via the command line
WEKA:: Interface
Uses flat text files to describe the data Can work with a wide variety of data files including its own .arff format and C4.5 file formats Data can be imported from a file in various formats:
Data can also be read from a URL or from an SQL database (using JDBC)
...
Pre-processing tools in WEKA are called filters WEKA contains filters for:
Classifiers in WEKA are models for predicting nominal or numeric quantities Implemented learning schemes include:
Decision trees and lists, instance-based classifiers, support vector machines, multi-layer perceptrons, logistic regression, Bayes nets, Bagging, boosting, stacking, error-correcting output codes, locally weighted learning,
Meta-classifiers include:
RapidMiner:: Introduction
intelligent data analysis, data mining, knowledge discovery, machine learning, predictive analytics, forecasting, and analytics in business intelligence (BI).
Is implemented in Java and available under GPL among other licenses Available from http://rapid-i.com
Is similar in spirit to Wekas Knowledge flow Data mining processes/routines are views as sequential operators
Operators define their expected inputs and delivered outputs as well as their parameters
Uses XML for describing operator trees in the KD process Alternatively can be started through the command line and passed the XML process file