8
Most read
14
Most read
20
Most read
GBM PACKAGE IN R
7/24/2014
Presentation Outline
• Algorithm Overview
• Basics
• How it solves problems
• Why to use it
• Deeper investigation while going through live code
What is GBM?
• Predictive modeling algorithm
• Classification & Regression
• Decision tree as a basis*
• Boosted
• Multiple weak models combined algorithmically
• Gradient boosted
• Iteratively solves residuals
• Stochastic
(some additional references on last slide)
* technically, GBM can take on other forms such as linear, but decision trees are the dominant usage,
Friedman specifically optimized for trees, and R’s implementation is internally represented as a tree
Predictive Modeling Landscape:
General Purpose Algorithms
(forillustrativepurposes only,nottoscale,precise,orcomprehensive;author’sperspective)
Linear Models Decision Trees Others
Linear Models
(lm)
Generalized
Linear Models (glm)
Regularized
Linear
Models
(glmnet)
Classification
And Regression
Trees (rpart)
Random
Forest
(randomForest)
Gradient
Boosted
Machines
(gbm)
Nearest Neighbor
(kNN)
Neural
Networks
(nnet)
Support
Vector
Machines
(kernlab)
complexity
Naïve Bayes
(klaR)
Splines
(earth)
More Comprehensive List: http://caret.r-forge.r-project.org/modelList.html
GBM’s decision tree structure
Why GBM?
• Characteristics
• Competitive Performance
• Robust
• Loss functions
• Fast (relatively)
• Usages
• Quick modeling
• Variable selection
• Final-stage precision modeling
Competitive Performance
• Competitive with high-end algorithms such as
RandomForest
• Reliable performance
• Avoids nonsensical predictions
• Rare to produce worse predictions than simpler models
• Often in winning Kaggle solutions
• Cited within winning solution descriptions in numerous
competitions, including $3M competition
• Many of the highest ranked competitors use it frequently
• Used in 4 of 5 personal top 20 finishes
Robust
• Explicitly handles NAs
• Scaling/normalization is unnecessary
• Handles more factor levels than random forest (1024 vs
32)
• Handles perfectly correlated independent variables
• No [known] limit to number of independent variables
Loss Functions
• Gaussian: squared loss
• Laplace: absolute loss
• Bernoulli: logistic, for 0/1
• Huberized: hinge, for 0/1
• Adaboost: exponential loss, for 0/1
• Multinomial: more than one class (produces probability matrix)
• Quantile: flexible alpha (e.g. optimize for 2 StDev threshold)
• Poisson: Poisson distribution, for counts
• CoxPH: Cox proportional hazard, for right-censored
• Tdist: t-distribution loss
• Pairwise: rankings (e.g. search result scoring)
• Concordant pairs
• Mean reciprocal rank
• Mean average precision
• Normalized discounted cumulative gain
Drawbacks
• Several hyper-parameters to tune
• I typically use roughly the same parameters to start, unless I
suspect the data set might have peculiar characteristics
• For creating a final model, tuning several parameters is advisable
• Still has capacity to overfit
• Despite internal cross-validation, it is still particularly prone to
overfit ID-like columns (suggestion: withhold them)
• Can have trouble with highly noisy data
• Black box
• However, GBM package does provide tools to analyze the resulting
models
Deeper Analysis via Walkthrough
• Hyper-parameter explanations (some, not all)
• Quickly analyze performance
• Analyze influence of variables
• Peek under the hood…then follow a toy problem
For those not attending the presentation, the code at the back is run at this
point and discussed. The remaining four slides were mainly to supplement
the discussion of the code and comments, and there was not sufficient
time.
Same analysis with a simpler data set
Note that one can recreate the
predictions of this first tree by finding the
terminal node for any prediction and
using the Prediction value (final column
in data frame). Those values for all
desired trees, plus the initial value (mean
for this) is the prediction.
Matches predictions 1 & 3
Matches predictions 2,4 & 5
Same analysis with a simpler data set
Explanation
1 tree built.
Tree has one decision only, node 0.
Node 0 indicates it split the 3rd field (SplitVar:2), to where values below 1.5
(ordered values 0 & 1 which are a & b) went to node 1; values above
1.5 (2/3 = c/d) went to node 2; missing (none) go to node 3.
Node 1 (X3=A/B) is a terminal node (SplitVar -1) and it predicts the mean plus -0.925.
Node 2 (X3=C/D) is a terminal node and it predicts the mean plus 1.01.
Node 3 (none) is a terminal node and it predicts the mean plus 0, effectively.
Later saw that gbm1$initF will show the intercept, which in this case is the mean.
GBM predict: fit a GBM to data
• gbm(formula = formula(data),
• distribution = "bernoulli",
• n.trees = 100,
• interaction.depth = 1,
• n.minobsinnode = 10,
• shrinkage = 0.001,
• bag.fraction = 0.5,
• train.fraction = 1.0,
• cv.folds=0,
• weights,
• data = list(),
• var.monotone = NULL,
• keep.data = TRUE,
• verbose = "CV",
• class.stratify.cv=NULL,
• n.cores = NULL)
Effect of shrinkage & trees
Source: https://www.youtube.com/watch?v=IXZKgIsZRm0 (GBM explanation by SciKit author)
Code Dump
• The code has been copied from a text R script into PowerPoint, so
the format isn’t great, but it should look OK if copying and pasting
back out to a text file. If not, here it is on Github.
• The code shown uses a competition data set that is comparable to
real world data and uses a simple GBM to predict sale prices of
construction equipment at auction.
• A GBM model was fit against 100k rows with 45-50 variables in about
2-4 minutes during the presentation. It improves the RMSE of
prediction against the mean from ~24.5k to ~9.7k, when scored on
data the model had not seen (and future dates, so the 100k/50k splits
should be valid), with fairly stable train:test performance.
• After predictions are made and scored, some GBM utilities are used
to see which variables the model found most influential, see how the
top 2 variables are used (per factor for one; throughout a continuous
distribution for the other), and see interaction effects of specific
variable pairs.
• Note: GBM was used by my teammate and I to finish 12th out of 476
in this competition (albeit a complex ensemble of GBMs)
Code Dump: Page1
library(Metrics) ##load evaluation package
setwd("C:/Users/Mark_Landry/Documents/K/dozer/")
##Done in advance to speed up loading of data set
train<-read.csv("Train.csv")
## Kaggle data set: http://www.kaggle.com/c/bluebook-for-bulldozers/data
train$saleTransform<-strptime(train$saledate,"%m/%d/%Y %H:%M")
train<-train[order(train$saleTransform),]
save(train,file="rTrain.Rdata")
load("rTrain.Rdata")
xTrain<-train[(nrow(train)-149999):(nrow(train)-50000),5:ncol(train)]
xTest<-train[(nrow(train)-49999):nrow(train),5:ncol(train)]
yTrain<-train[(nrow(train)-149999):(nrow(train)-50000),2]
yTest<-train[(nrow(train)-49999):nrow(train),2]
dim(xTrain); dim(xTest)
sapply(xTrain,function(x) length(levels(x)))
## check levels; gbm is robust, but still has a limit of 1024 per factor; for initial model, remove
## after iterating through model, would want to go back and compress these factors to investigate
## their usefulness (or other information analysis)
xTrain$saledate<-NULL; xTest$saledate<-NULL
xTrain$fiModelDesc<-NULL; xTest$fiModelDesc<-NULL
xTrain$fiBaseModel<-NULL; xTest$fiBaseModel<-NULL
xTrain$saleTransform<-NULL; xTest$saleTransform<-NULL
Code Dump: Page2
library(gbm)
## Set up parameters to pass in; there are many more hyper-parameters available, but these are the most common to
control
GBM_NTREES = 400
## 400 trees in the model; can scale back later for predictions, if desired or overfitting is suspected
GBM_SHRINKAGE = 0.05
## shrinkage is a regularization parameter dictating how fast/aggressive the algorithm moves across
the loss gradient
## 0.05 is somewhat aggressive; default is 0.001, values below 0.1 tend to produce good results
## decreasing shrinkage generally improves results, but requires more trees, so the two
should be adjusted in tandem
GBM_DEPTH = 4
## depth 4 means each tree will evaluate four decisions;
## will always yield [3*depth + 1] nodes and [2*depth + 1] terminal nodes (depth 4 = 9)
## because each decision yields 3 nodes, but each decision will come from a prior node
GBM_MINOBS = 30
## regularization parameter to dictate how many observations must be present to yield a terminal node
## higher number means more conservative fit; 30 is fairly high, but good for exploratory fits; default is
10
## Fit model
g<-gbm.fit(x=xTrain,y=yTrain,distribution = "gaussian",n.trees = GBM_NTREES,shrinkage = GBM_SHRINKAGE,
interaction.depth = GBM_DEPTH,n.minobsinnode = GBM_MINOBS)
## gbm fit; provide all remaining independent variables in xTrain; provide targets as yTrain;
## gaussian distribution will optimize squared loss;
Code Dump: Page3
## get predictions; first on train set, then on unseen test data
tP1 <- predict.gbm(object = g,newdata = xTrain,GBM_NTREES)
hP1 <- predict.gbm(object = g,newdata = xTest,GBM_NTREES)
## compare model performance to default (overall mean)
rmse(yTrain,tP1) ## 9452.742 on data used for training
rmse(yTest,hP1) ## 9740.559 ~3% drop on unseen data; does not seem
to be overfit
rmse(yTest,mean(yTrain)) ## 24481.08 overall mean; cut error rate (from perfection) by 60%
## look at variables
summary(g) ## summary will plot and then show the relative influence of each variable to the entire GBM model (all trees)
## test dominant variable mean
library(sqldf)
trainProdClass<-as.data.frame(cbind(as.character(xTrain$fiProductClassDesc),yTrain))
testProdClass<-as.data.frame(cbind(as.character(xTest$fiProductClassDesc),yTest))
colnames(trainProdClass)<-c("fiProductClassDesc","y"); colnames(testProdClass)<-c("fiProductClassDesc","y")
ProdClassMeans<-sqldf("SELECT fiProductClassDesc,avg(y) avg, COUNT(*) n FROM trainProdClass GROUP BY
fiProductClassDesc")
ProdClassPredictions<-sqldf("SELECT case when n > 30 then avg ELSE 31348.63 end avg
FROM ProdClassMeans P LEFT JOIN testProdClass t ON t.fiProductClassDesc = P.fiProductClassDesc")
rmse(yTest,ProdClassPredictions$avg) ## 29082.64 ? peculiar result on the fiProductClassDesc means, which seemed
fairly stable and useful
##seems to say that the primary factor alone is not helpful; full tree needed
Code Dump: Page4
## Investigate actual GBM model
pretty.gbm.tree(g,1) ## show underlying model for the first decision tree
summary(xTrain[,10]) ## underlying model showed variable 9 to be first point in tree (9 with 0 index = 10th
column)
g$initF ## view what is effectively the "y intercept"
mean(yTrain) ## equivalence shows gaussian y intercept is the mean
t(g$c.splits[1][[1]]) ## show whether each factor level should go left or right
plot(g,10) ## plot fiProductClassDesc, the variable with the highest
rel.inf
plot(g,3) ## plot YearMade, continuous variable with 2nd highest
rel.inf
interact.gbm(g,xTrain,c(10,3))
## compute H statistic to
show interaction; integrates
interact.gbm(g,xTrain,c(10,3))
## example of uninteresting
interaction
Selected References
• CRAN
• Documentation
• vignette
• Algorithm publications:
• Greedy function approximation: A gradient boosting machine
Friedman 2/99
• Stochastic Gradient Boosting; Friedman 3/99
• Overviews
• Gradient boosting machines, a tutorial: Frontiers (4/13)
• Wikipedia (pretty good article, really)
• Video of author of GBM in Python: Gradient Boosted Regression
Trees in scikit-learn
• Very helpful, but the implementation is not decision “stumps” in R, so
some things are different in R (e.g. number of trees need not be so high)

More Related Content

PDF
Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation
PPTX
Digital Image Communication and Enhancement
PPT
Fuzzy c-means clustering for image segmentation
PPTX
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8
PPTX
Feature selection concepts and methods
PPTX
PDF
Dr. Trevor Hastie: Data Science of GBM (October 10, 2013: Presented With H2O)
PDF
Decision trees in Machine Learning
Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation
Digital Image Communication and Enhancement
Fuzzy c-means clustering for image segmentation
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8
Feature selection concepts and methods
Dr. Trevor Hastie: Data Science of GBM (October 10, 2013: Presented With H2O)
Decision trees in Machine Learning

What's hot (20)

PPTX
Optimization/Gradient Descent
PPTX
KNN Classifier
PDF
Genetic Algorithm for optimization on IRIS Dataset presentation ppt
PDF
Feature selection
PPT
Bayes Classification
PPTX
Decision trees and random forests
PDF
Alternating direction
PPT
Binomial Heaps
PDF
Understanding random forests
PPTX
Analytic hierarchy process
PPTX
Lecture 18: Gaussian Mixture Models and Expectation Maximization
PDF
Big m method
PDF
Deep Feed Forward Neural Networks and Regularization
PPT
Chapter 5
DOCX
Stock market trend prediction using k nearest neighbor(knn) algorithm
PPT
K means Clustering Algorithm
PDF
Decision tree
PPT
Decision tree
PDF
Gradient boosting in practice: a deep dive into xgboost
Optimization/Gradient Descent
KNN Classifier
Genetic Algorithm for optimization on IRIS Dataset presentation ppt
Feature selection
Bayes Classification
Decision trees and random forests
Alternating direction
Binomial Heaps
Understanding random forests
Analytic hierarchy process
Lecture 18: Gaussian Mixture Models and Expectation Maximization
Big m method
Deep Feed Forward Neural Networks and Regularization
Chapter 5
Stock market trend prediction using k nearest neighbor(knn) algorithm
K means Clustering Algorithm
Decision tree
Decision tree
Gradient boosting in practice: a deep dive into xgboost
Ad

Viewers also liked (8)

PPTX
Inlining Heuristics
PPTX
Gbm.more GBM in H2O
PPTX
XGBoost (System Overview)
PPTX
Automated data analysis with Python
PDF
GBM theory code and parameters
PDF
Kaggle Winning Solution Xgboost algorithm -- Let us learn from its author
PDF
PDF
Decision Tree Ensembles - Bagging, Random Forest & Gradient Boosting Machines
Inlining Heuristics
Gbm.more GBM in H2O
XGBoost (System Overview)
Automated data analysis with Python
GBM theory code and parameters
Kaggle Winning Solution Xgboost algorithm -- Let us learn from its author
Decision Tree Ensembles - Bagging, Random Forest & Gradient Boosting Machines
Ad

Similar to GBM package in r (20)

PPTX
Cubesat challenge considerations deep dive
PPTX
Algorithm explanations
PPTX
Machine Learning Algorithms (Part 1)
PPT
design mapping lecture6-mapreducealgorithmdesign.ppt
PDF
Machine Learning - Supervised Learning
PPTX
greedy algorithm.pptx good for understanding
PDF
Genetic programming
PPTX
XgBoost.pptx
PDF
MapReduce Algorithm Design - Parallel Reduce Operations
PPTX
Minmax and alpha beta pruning.pptx
PPTX
Kaggle review Planet: Understanding the Amazon from Space
PPTX
Musings of kaggler
PPTX
ngboost.pptx
PPTX
CLUSTER ANALYSIS ALGORITHMS.pptx
PDF
Heuristic design of experiments w meta gradient search
PDF
Firefly exact MCMC for Big Data
PDF
Tensors Are All You Need: Faster Inference with Hummingbird
PPTX
Machine Learning with Python unit-2.pptx
PDF
R user group meeting 25th jan 2017
PPTX
Decision tree induction
Cubesat challenge considerations deep dive
Algorithm explanations
Machine Learning Algorithms (Part 1)
design mapping lecture6-mapreducealgorithmdesign.ppt
Machine Learning - Supervised Learning
greedy algorithm.pptx good for understanding
Genetic programming
XgBoost.pptx
MapReduce Algorithm Design - Parallel Reduce Operations
Minmax and alpha beta pruning.pptx
Kaggle review Planet: Understanding the Amazon from Space
Musings of kaggler
ngboost.pptx
CLUSTER ANALYSIS ALGORITHMS.pptx
Heuristic design of experiments w meta gradient search
Firefly exact MCMC for Big Data
Tensors Are All You Need: Faster Inference with Hummingbird
Machine Learning with Python unit-2.pptx
R user group meeting 25th jan 2017
Decision tree induction

Recently uploaded (20)

PPTX
sac 451hinhgsgshssjsjsjheegdggeegegdggddgeg.pptx
PDF
Navigating the Thai Supplements Landscape.pdf
PPTX
Phase1_final PPTuwhefoegfohwfoiehfoegg.pptx
PPT
DU, AIS, Big Data and Data Analytics.ppt
PPTX
chrmotography.pptx food anaylysis techni
PPTX
DS-40-Pre-Engagement and Kickoff deck - v8.0.pptx
PPTX
ai agent creaction with langgraph_presentation_
PPTX
AI AND ML PROPOSAL PRESENTATION MUST.pptx
PPTX
Tapan_20220802057_Researchinternship_final_stage.pptx
PPTX
CYBER SECURITY the Next Warefare Tactics
PPTX
New ISO 27001_2022 standard and the changes
PPT
lectureusjsjdhdsjjshdshshddhdhddhhd1.ppt
PPT
statistics analysis - topic 3 - describing data visually
PDF
OneRead_20250728_1808.pdfhdhddhshahwhwwjjaaja
PDF
ahaaaa shbzjs yaiw jsvssv bdjsjss shsusus s
PDF
©️ 02_SKU Automatic SW Robotics for Microsoft PC.pdf
PPTX
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
PPTX
recommendation Project PPT with details attached
PPTX
statsppt this is statistics ppt for giving knowledge about this topic
PPTX
CHAPTER-2-THE-ACCOUNTING-PROCESS-2-4.pptx
sac 451hinhgsgshssjsjsjheegdggeegegdggddgeg.pptx
Navigating the Thai Supplements Landscape.pdf
Phase1_final PPTuwhefoegfohwfoiehfoegg.pptx
DU, AIS, Big Data and Data Analytics.ppt
chrmotography.pptx food anaylysis techni
DS-40-Pre-Engagement and Kickoff deck - v8.0.pptx
ai agent creaction with langgraph_presentation_
AI AND ML PROPOSAL PRESENTATION MUST.pptx
Tapan_20220802057_Researchinternship_final_stage.pptx
CYBER SECURITY the Next Warefare Tactics
New ISO 27001_2022 standard and the changes
lectureusjsjdhdsjjshdshshddhdhddhhd1.ppt
statistics analysis - topic 3 - describing data visually
OneRead_20250728_1808.pdfhdhddhshahwhwwjjaaja
ahaaaa shbzjs yaiw jsvssv bdjsjss shsusus s
©️ 02_SKU Automatic SW Robotics for Microsoft PC.pdf
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
recommendation Project PPT with details attached
statsppt this is statistics ppt for giving knowledge about this topic
CHAPTER-2-THE-ACCOUNTING-PROCESS-2-4.pptx

GBM package in r

  • 1. GBM PACKAGE IN R 7/24/2014
  • 2. Presentation Outline • Algorithm Overview • Basics • How it solves problems • Why to use it • Deeper investigation while going through live code
  • 3. What is GBM? • Predictive modeling algorithm • Classification & Regression • Decision tree as a basis* • Boosted • Multiple weak models combined algorithmically • Gradient boosted • Iteratively solves residuals • Stochastic (some additional references on last slide) * technically, GBM can take on other forms such as linear, but decision trees are the dominant usage, Friedman specifically optimized for trees, and R’s implementation is internally represented as a tree
  • 4. Predictive Modeling Landscape: General Purpose Algorithms (forillustrativepurposes only,nottoscale,precise,orcomprehensive;author’sperspective) Linear Models Decision Trees Others Linear Models (lm) Generalized Linear Models (glm) Regularized Linear Models (glmnet) Classification And Regression Trees (rpart) Random Forest (randomForest) Gradient Boosted Machines (gbm) Nearest Neighbor (kNN) Neural Networks (nnet) Support Vector Machines (kernlab) complexity Naïve Bayes (klaR) Splines (earth) More Comprehensive List: http://caret.r-forge.r-project.org/modelList.html
  • 6. Why GBM? • Characteristics • Competitive Performance • Robust • Loss functions • Fast (relatively) • Usages • Quick modeling • Variable selection • Final-stage precision modeling
  • 7. Competitive Performance • Competitive with high-end algorithms such as RandomForest • Reliable performance • Avoids nonsensical predictions • Rare to produce worse predictions than simpler models • Often in winning Kaggle solutions • Cited within winning solution descriptions in numerous competitions, including $3M competition • Many of the highest ranked competitors use it frequently • Used in 4 of 5 personal top 20 finishes
  • 8. Robust • Explicitly handles NAs • Scaling/normalization is unnecessary • Handles more factor levels than random forest (1024 vs 32) • Handles perfectly correlated independent variables • No [known] limit to number of independent variables
  • 9. Loss Functions • Gaussian: squared loss • Laplace: absolute loss • Bernoulli: logistic, for 0/1 • Huberized: hinge, for 0/1 • Adaboost: exponential loss, for 0/1 • Multinomial: more than one class (produces probability matrix) • Quantile: flexible alpha (e.g. optimize for 2 StDev threshold) • Poisson: Poisson distribution, for counts • CoxPH: Cox proportional hazard, for right-censored • Tdist: t-distribution loss • Pairwise: rankings (e.g. search result scoring) • Concordant pairs • Mean reciprocal rank • Mean average precision • Normalized discounted cumulative gain
  • 10. Drawbacks • Several hyper-parameters to tune • I typically use roughly the same parameters to start, unless I suspect the data set might have peculiar characteristics • For creating a final model, tuning several parameters is advisable • Still has capacity to overfit • Despite internal cross-validation, it is still particularly prone to overfit ID-like columns (suggestion: withhold them) • Can have trouble with highly noisy data • Black box • However, GBM package does provide tools to analyze the resulting models
  • 11. Deeper Analysis via Walkthrough • Hyper-parameter explanations (some, not all) • Quickly analyze performance • Analyze influence of variables • Peek under the hood…then follow a toy problem For those not attending the presentation, the code at the back is run at this point and discussed. The remaining four slides were mainly to supplement the discussion of the code and comments, and there was not sufficient time.
  • 12. Same analysis with a simpler data set Note that one can recreate the predictions of this first tree by finding the terminal node for any prediction and using the Prediction value (final column in data frame). Those values for all desired trees, plus the initial value (mean for this) is the prediction. Matches predictions 1 & 3 Matches predictions 2,4 & 5
  • 13. Same analysis with a simpler data set Explanation 1 tree built. Tree has one decision only, node 0. Node 0 indicates it split the 3rd field (SplitVar:2), to where values below 1.5 (ordered values 0 & 1 which are a & b) went to node 1; values above 1.5 (2/3 = c/d) went to node 2; missing (none) go to node 3. Node 1 (X3=A/B) is a terminal node (SplitVar -1) and it predicts the mean plus -0.925. Node 2 (X3=C/D) is a terminal node and it predicts the mean plus 1.01. Node 3 (none) is a terminal node and it predicts the mean plus 0, effectively. Later saw that gbm1$initF will show the intercept, which in this case is the mean.
  • 14. GBM predict: fit a GBM to data • gbm(formula = formula(data), • distribution = "bernoulli", • n.trees = 100, • interaction.depth = 1, • n.minobsinnode = 10, • shrinkage = 0.001, • bag.fraction = 0.5, • train.fraction = 1.0, • cv.folds=0, • weights, • data = list(), • var.monotone = NULL, • keep.data = TRUE, • verbose = "CV", • class.stratify.cv=NULL, • n.cores = NULL)
  • 15. Effect of shrinkage & trees Source: https://www.youtube.com/watch?v=IXZKgIsZRm0 (GBM explanation by SciKit author)
  • 16. Code Dump • The code has been copied from a text R script into PowerPoint, so the format isn’t great, but it should look OK if copying and pasting back out to a text file. If not, here it is on Github. • The code shown uses a competition data set that is comparable to real world data and uses a simple GBM to predict sale prices of construction equipment at auction. • A GBM model was fit against 100k rows with 45-50 variables in about 2-4 minutes during the presentation. It improves the RMSE of prediction against the mean from ~24.5k to ~9.7k, when scored on data the model had not seen (and future dates, so the 100k/50k splits should be valid), with fairly stable train:test performance. • After predictions are made and scored, some GBM utilities are used to see which variables the model found most influential, see how the top 2 variables are used (per factor for one; throughout a continuous distribution for the other), and see interaction effects of specific variable pairs. • Note: GBM was used by my teammate and I to finish 12th out of 476 in this competition (albeit a complex ensemble of GBMs)
  • 17. Code Dump: Page1 library(Metrics) ##load evaluation package setwd("C:/Users/Mark_Landry/Documents/K/dozer/") ##Done in advance to speed up loading of data set train<-read.csv("Train.csv") ## Kaggle data set: http://www.kaggle.com/c/bluebook-for-bulldozers/data train$saleTransform<-strptime(train$saledate,"%m/%d/%Y %H:%M") train<-train[order(train$saleTransform),] save(train,file="rTrain.Rdata") load("rTrain.Rdata") xTrain<-train[(nrow(train)-149999):(nrow(train)-50000),5:ncol(train)] xTest<-train[(nrow(train)-49999):nrow(train),5:ncol(train)] yTrain<-train[(nrow(train)-149999):(nrow(train)-50000),2] yTest<-train[(nrow(train)-49999):nrow(train),2] dim(xTrain); dim(xTest) sapply(xTrain,function(x) length(levels(x))) ## check levels; gbm is robust, but still has a limit of 1024 per factor; for initial model, remove ## after iterating through model, would want to go back and compress these factors to investigate ## their usefulness (or other information analysis) xTrain$saledate<-NULL; xTest$saledate<-NULL xTrain$fiModelDesc<-NULL; xTest$fiModelDesc<-NULL xTrain$fiBaseModel<-NULL; xTest$fiBaseModel<-NULL xTrain$saleTransform<-NULL; xTest$saleTransform<-NULL
  • 18. Code Dump: Page2 library(gbm) ## Set up parameters to pass in; there are many more hyper-parameters available, but these are the most common to control GBM_NTREES = 400 ## 400 trees in the model; can scale back later for predictions, if desired or overfitting is suspected GBM_SHRINKAGE = 0.05 ## shrinkage is a regularization parameter dictating how fast/aggressive the algorithm moves across the loss gradient ## 0.05 is somewhat aggressive; default is 0.001, values below 0.1 tend to produce good results ## decreasing shrinkage generally improves results, but requires more trees, so the two should be adjusted in tandem GBM_DEPTH = 4 ## depth 4 means each tree will evaluate four decisions; ## will always yield [3*depth + 1] nodes and [2*depth + 1] terminal nodes (depth 4 = 9) ## because each decision yields 3 nodes, but each decision will come from a prior node GBM_MINOBS = 30 ## regularization parameter to dictate how many observations must be present to yield a terminal node ## higher number means more conservative fit; 30 is fairly high, but good for exploratory fits; default is 10 ## Fit model g<-gbm.fit(x=xTrain,y=yTrain,distribution = "gaussian",n.trees = GBM_NTREES,shrinkage = GBM_SHRINKAGE, interaction.depth = GBM_DEPTH,n.minobsinnode = GBM_MINOBS) ## gbm fit; provide all remaining independent variables in xTrain; provide targets as yTrain; ## gaussian distribution will optimize squared loss;
  • 19. Code Dump: Page3 ## get predictions; first on train set, then on unseen test data tP1 <- predict.gbm(object = g,newdata = xTrain,GBM_NTREES) hP1 <- predict.gbm(object = g,newdata = xTest,GBM_NTREES) ## compare model performance to default (overall mean) rmse(yTrain,tP1) ## 9452.742 on data used for training rmse(yTest,hP1) ## 9740.559 ~3% drop on unseen data; does not seem to be overfit rmse(yTest,mean(yTrain)) ## 24481.08 overall mean; cut error rate (from perfection) by 60% ## look at variables summary(g) ## summary will plot and then show the relative influence of each variable to the entire GBM model (all trees) ## test dominant variable mean library(sqldf) trainProdClass<-as.data.frame(cbind(as.character(xTrain$fiProductClassDesc),yTrain)) testProdClass<-as.data.frame(cbind(as.character(xTest$fiProductClassDesc),yTest)) colnames(trainProdClass)<-c("fiProductClassDesc","y"); colnames(testProdClass)<-c("fiProductClassDesc","y") ProdClassMeans<-sqldf("SELECT fiProductClassDesc,avg(y) avg, COUNT(*) n FROM trainProdClass GROUP BY fiProductClassDesc") ProdClassPredictions<-sqldf("SELECT case when n > 30 then avg ELSE 31348.63 end avg FROM ProdClassMeans P LEFT JOIN testProdClass t ON t.fiProductClassDesc = P.fiProductClassDesc") rmse(yTest,ProdClassPredictions$avg) ## 29082.64 ? peculiar result on the fiProductClassDesc means, which seemed fairly stable and useful ##seems to say that the primary factor alone is not helpful; full tree needed
  • 20. Code Dump: Page4 ## Investigate actual GBM model pretty.gbm.tree(g,1) ## show underlying model for the first decision tree summary(xTrain[,10]) ## underlying model showed variable 9 to be first point in tree (9 with 0 index = 10th column) g$initF ## view what is effectively the "y intercept" mean(yTrain) ## equivalence shows gaussian y intercept is the mean t(g$c.splits[1][[1]]) ## show whether each factor level should go left or right plot(g,10) ## plot fiProductClassDesc, the variable with the highest rel.inf plot(g,3) ## plot YearMade, continuous variable with 2nd highest rel.inf interact.gbm(g,xTrain,c(10,3)) ## compute H statistic to show interaction; integrates interact.gbm(g,xTrain,c(10,3)) ## example of uninteresting interaction
  • 21. Selected References • CRAN • Documentation • vignette • Algorithm publications: • Greedy function approximation: A gradient boosting machine Friedman 2/99 • Stochastic Gradient Boosting; Friedman 3/99 • Overviews • Gradient boosting machines, a tutorial: Frontiers (4/13) • Wikipedia (pretty good article, really) • Video of author of GBM in Python: Gradient Boosted Regression Trees in scikit-learn • Very helpful, but the implementation is not decision “stumps” in R, so some things are different in R (e.g. number of trees need not be so high)