How to build Ensemble Models in machine learning? (with code in R)

sauravkaushik8 25 Jun, 2019
12 min read

Introduction

Over the last 12 months, I have been participating in a number of machine learning hackathons on Analytics Vidhya and Kaggle competitions. After the competition, I always make sure to go through the winner’s solution. The winner’s solution usually provide me critical insights, which have helped me immensely in future competitions.

Most of the winners rely on an ensemble of well-tuned individual models along with feature engineering. If you are starting with machine learning, I would advise you to lay emphasis on these two areas as I have found them equally important to do well in a machine learning.

Most of the time, I was able to crack the feature engineering part but probably didn’t use the ensemble of multiple models. If you are a beginner, it’s even better to get familiar with ensembling as early as possible. Chances are that you are already applying it without knowing!

In this article, I’ll take you through the basics of ensemble modeling. Then I will walk you through the advantages of ensembling. Also, to provide you hands-on experience on ensemble modeling, we will use ensembling on a hackathon problem using R.

 

 

Table of Content

  1. What is ensembling?
  2. Types of ensembling
  3. Advantages and disadvantages of ensembling
  4. Practical guide to implementing ensembling in R
  5. Additional resources

P.S. For this article, we will assume that you can build individual models in R / Python. If not, you can start your journey with our learning path.

 

1.What is ensembling?

In general, ensembling is a technique of combining two or more algorithms of similar or dissimilar types called base learners. This is done to make a more robust system which incorporates the predictions from all the base learners. It can be understood as conference room meeting between multiple traders to make a decision on whether the price of a stock will go up or not.

Since all of them have a different understanding of the stock market and thus a different mapping function from the problem statement to the desired outcome. Therefore, they are supposed to make varied predictions on the stock price based on their own understandings of the market.

Now we can take all of these predictions into account while making the final decision. This will make our final decision more robust, accurate and less likely to be biased. The final decision would have been opposite if one of these traders would have made this decision alone.

You can consider another example of a candidate going through multiple rounds of job interviews. The final decision of candidate’s ability is generally taken based on the feedback of all the interviewers. Although a single interviewer might not be able to test the candidate for each required skill and trait. But the combined feedback of multiple interviewers usually helps in better assessment of the candidate.

 

2. Types of ensembling

Some of the basic concepts which you should be aware of before we go into further detail are:

  • Averaging: It’s defined as taking the average of predictions from models in case of regression problem or while predicting probabilities for the classification problem.
  • Majority vote: It’s defined as taking the prediction with maximum vote / recommendation from multiple models predictions while predicting the outcomes of a classification problem.
  • Weighted average: In this, different weights are applied to predictions from multiple models then taking the average which means giving high or low importance to specific model output.

 

Practically speaking, there can be a countless number of ways in which you can ensemble different models. But these are some techniques that are mostly used:

  1. Bagging: Bagging is also referred to as bootstrap aggregation. To understand bagging, we first need to understand bootstrapping. Bootstrapping is a sampling technique in which we choose ‘n’ observations or rows out of the original dataset of ‘n’ rows as well. But the key is that each row is selected with replacement from the original dataset so that each row is equally likely to be selected in each iteration. Let’s say we have 3 rows numbered 1, 2 and 3.

For bootstrapped sample, we choose one out of these three randomly. Say we chose Row 2.

You see that even though Row 2 is chosen from the data to the bootstrap sample, it’s still present in the data. Now, each of the three:

Rows have the same probability of being selected again. Let’s say we choose Row 1 this time.

Again, each row in the data has the same probability to be chosen for Bootstrapped sample. Let’s say we randomly choose Row 1 again.

Thus, we can have multiple bootstrapped samples from the same data. Once we have these multiple bootstrapped samples, we can grow trees for each of these bootstrapped samples and use the majority vote or averaging concepts to get the final prediction. This is how bagging works.

One important thing to note here is that it’s done mainly to reduce the variance. Now, random forest actually uses this concept but it goes a step ahead to further reduce the variance by randomly choosing a subset of features as well for each bootstrapped sample to make the splits while training.

  1. Boosting: Boosting is a sequential technique in which, the first algorithm is trained on the entire dataset and the subsequent algorithms are built by fitting the residuals of the first algorithm, thus giving higher weight to those observations that were poorly predicted by the previous model.

It relies on creating a series of weak learners each of which might not be good for the entire dataset but is good for some part of the dataset. Thus, each model actually boosts the performance of the ensemble.

It’s really important to note that boosting is focused on reducing the bias. This makes the boosting algorithms prone to overfitting. Thus, parameter tuning becomes a crucial part of boosting algorithms to make them avoid overfitting.

Some examples of boosting are XGBoost, GBM, ADABOOST, etc.

  1. Stacking: In stacking multiple layers of machine learning models are placed one over another where each of the models passes their predictions to the model in the layer above it and the top layer model takes decisions based on the outputs of the models in layers below it.

Let’s understand it with an example:

Here, we have two layers of machine learning models:

  • Bottom layer models (d1, d2, d3 ) which receive the original input features(x) from the dataset.
  • Top layer model, f() which takes the output of the bottom layer models (d1, d2, d3 ) as its input and predicts the final output.
  • One key thing to note here is that out of fold predictions are used while predicting for the training data.

Here, we have used only two layers but it can be any number of layers and any number of models in each layer. Two of the key principles for selecting the models:

  • The individual models fulfill particular accuracy criteria.
  • The model predictions of various individual models are not highly correlated with the predictions of other models.

One thing that you might have realized is that we have used the top layer model which takes as input the predictions of the bottom layer models. This top layer model can also be replaced by many other simpler formulas like:

  • Averaging 
  • Majority vote 
  • Weighted average

 

3. Advantages and Disadvantages of ensembling

3.1 Advantages

  • Ensembling is a proven method for improving the accuracy of the model and works in most of the cases.
  • It is the key ingredient for winning almost all of the machine learning hackathons.
  • Ensembling makes the model more robust and stable thus ensuring decent performance on the test cases in most scenarios.
  • You can use ensembling to capture linear and simple as well non-linear complex relationships in the data. This can be done by using two different models and forming an ensemble of two.

 

3.2 Disadvantages

  • Ensembling reduces the model interpretability and makes it very difficult to draw any crucial business insights at the end.
  • It is time-consuming and thus might not be the best idea for real-time applications.
  • The selection of models for creating an ensemble is an art which is really hard to master.

 

4. Practical guide to implementing ensembling in R

I believe you would have a good grasp on ensembling concepts by now. Well, enough of theory now, let’s get down to implementing ensembling and see whether it can help us improve our accuracy for a real machine learning challenge. If you wish to read more about the basics of ensembling, then you can refer to this resource.

For the purpose of implementing ensembling, I have chosen Loan Prediction problem. We have to predict whether the bank should approve the loan based on the applicant profile or not. It’s a binary classification problem. You can read more about the problem here.

I’ll be using caret package in R for training various individual models. It’s the goto package for modeling in R. Don’t worry if you are not familiar with the caret package, you can get through this article to get the comprehensive knowledge of caret package. Let’s get done with getting the data and data cleaning part.

#Loading the required libraries
library('caret')
#Seeting the random seed
set.seed(1)

#Loading the hackathon dataset
data<-read.csv(url('https://datahack-prod.s3.ap-south-1.amazonaws.com/train_file/train_u6lujuX_CVtuZ9i.csv'))

#Let's see if the structure of dataset data
str(data)
'data.frame':            614 obs. of  13 variables:
$ Loan_ID          : Factor w/ 614 levels "LP001002","LP001003",..: 1 2 3 4 5 6 7 8 9 10 ...
$ Gender           : Factor w/ 3 levels "","Female","Male": 3 3 3 3 3 3 3 3 3 3 ...
$ Married          : Factor w/ 3 levels "","No","Yes": 2 3 3 3 2 3 3 3 3 3 ...
$ Dependents       : Factor w/ 5 levels "","0","1","2",..: 2 3 2 2 2 4 2 5 4 3 ...
$ Education        : Factor w/ 2 levels "Graduate","Not Graduate": 1 1 1 2 1 1 2 1 1 1 ...
$ Self_Employed    : Factor w/ 3 levels "","No","Yes": 2 2 3 2 2 3 2 2 2 2 ...
$ ApplicantIncome  : int  5849 4583 3000 2583 6000 5417 2333 3036 4006 12841 ...
$ CoapplicantIncome: num  0 1508 0 2358 0 ...
$ LoanAmount       : int  NA 128 66 120 141 267 95 158 168 349 ...
$ Loan_Amount_Term : int  360 360 360 360 360 360 360 360 360 360 ...
$ Credit_History   : int  1 1 1 1 1 1 1 0 1 1 ...
$ Property_Area    : Factor w/ 3 levels "Rural","Semiurban",..: 3 1 3 3 3 3 3 2 3 2 ...
$ Loan_Status      : Factor w/ 2 levels "N","Y": 2 1 2 2 2 2 2 1 2 1 ...

#Does the data contain missing values
sum(is.na(data))
[1] 86

#Imputing missing values using median
preProcValues <- preProcess(data, method = c("medianImpute","center","scale"))
library('RANN')
data_processed <- predict(preProcValues, data)

sum(is.na(data_processed))
[1] 0
#Spliting training set into two parts based on outcome: 75% and 25%
index <- createDataPartition(data_processed$Loan_Status, p=0.75, list=FALSE)
trainSet <- data_processed[ index,]
testSet <- data_processed[-index,]

I have divided the data into two parts which I’ll be using to simulate the training and testing operations. We now define the training controls and the predictor and outcome variables:

#Defining the training controls for multiple models
fitControl <- trainControl(
  method = "cv",
  number = 5,
savePredictions = 'final',
classProbs = T)

#Defining the predictors and outcome
predictors<-c("Credit_History", "LoanAmount", "Loan_Amount_Term", "ApplicantIncome",
  "CoapplicantIncome")
outcomeName<-'Loan_Status'

Now let’s get started with training a random forest and test its accuracy on the test set that we have created:

#Training the random forest model
model_rf<-train(trainSet[,predictors],trainSet[,outcomeName],method='rf',trControl=fitControl,tuneLength=3)

#Predicting using random forest model
testSet$pred_rf<-predict(object = model_rf,testSet[,predictors])

#Checking the accuracy of the random forest model
confusionMatrix(testSet$Loan_Status,testSet$pred_rf)

Confusion Matrix and Statistics
Reference
Prediction  N  Y
N 28 20
Y  9 96
Accuracy : 0.8105         
95% CI : (0.7393, 0.8692)
No Information Rate : 0.7582         
P-Value [Acc > NIR] : 0.07566        
Kappa : 0.5306         
Mcnemar's Test P-Value : 0.06332        
Sensitivity : 0.7568         
Specificity : 0.8276         
Pos Pred Value : 0.5833         
Neg Pred Value : 0.9143         
Prevalence : 0.2418         
Detection Rate : 0.1830         
Detection Prevalence : 0.3137         
Balanced Accuracy : 0.7922         
'Positive' Class : N

Well, as you can see, we got 0.81 accuracy with the individual random forest model. Let’s see the performance of KNN:

#Training the knn model
model_knn<-train(trainSet[,predictors],trainSet[,outcomeName],method='knn',trControl=fitControl,tuneLength=3)

#Predicting using knn model
testSet$pred_knn<-predict(object = model_knn,testSet[,predictors])

#Checking the accuracy of the random forest model
confusionMatrix(testSet$Loan_Status,testSet$pred_knn)

Confusion Matrix and Statistics
Reference
Prediction   N   Y
N  29  19
Y   2 103
Accuracy : 0.8627        
95% CI : (0.7979, 0.913)
No Information Rate : 0.7974        
P-Value [Acc > NIR] : 0.0241694      
Kappa : 0.6473        
Mcnemar's Test P-Value : 0.0004803     
Sensitivity : 0.9355        
Specificity : 0.8443        
Pos Pred Value : 0.6042        
Neg Pred Value : 0.9810        
Prevalence : 0.2026        
Detection Rate : 0.1895        
Detection Prevalence : 0.3137        
Balanced Accuracy : 0.8899        
'Positive' Class : N

It’s great since we are able to get 0.86 accuracy with the individual KNN model. Let’s see the performance of Logistic regression as well before we go on to create ensemble of these three.

#Training the Logistic regression model
model_lr<-train(trainSet[,predictors],trainSet[,outcomeName],method='glm',trControl=fitControl,tuneLength=3)

#Predicting using knn model
testSet$pred_lr<-predict(object = model_lr,testSet[,predictors])

#Checking the accuracy of the random forest model
confusionMatrix(testSet$Loan_Status,testSet$pred_lr)

Confusion Matrix and Statistics
Reference
Prediction   N   Y
N  29  19
Y   2 103
Accuracy : 0.8627        
95% CI : (0.7979, 0.913)
No Information Rate : 0.7974        
P-Value [Acc > NIR] : 0.0241694     
Kappa : 0.6473        
Mcnemar's Test P-Value : 0.0004803     
Sensitivity : 0.9355        
Specificity : 0.8443        
Pos Pred Value : 0.6042        
Neg Pred Value : 0.9810        
Prevalence : 0.2026        
Detection Rate : 0.1895        
Detection Prevalence : 0.3137        
Balanced Accuracy : 0.8899        
'Positive' Class : N

And the logistic regression also gives us the accuracy of 0.86.

Now, let’s try out different ways of forming an ensemble with these models as we have discussed:

  • Averaging: In this, we’ll average the predictions from the three models. Since the predictions are either ‘Y’ or ‘N’, averaging doesn’t make much sense for this binary classification. However, we can do averaging on the probabilities of observations to be in wither of these binary classes.
#Predicting the probabilities
testSet$pred_rf_prob<-predict(object = model_rf,testSet[,predictors],type='prob')
testSet$pred_knn_prob<-predict(object = model_knn,testSet[,predictors],type='prob')
testSet$pred_lr_prob<-predict(object = model_lr,testSet[,predictors],type='prob')

#Taking average of predictions
testSet$pred_avg<-(testSet$pred_rf_prob$Y+testSet$pred_knn_prob$Y+testSet$pred_lr_prob$Y)/3

#Splitting into binary classes at 0.5
testSet$pred_avg<-as.factor(ifelse(testSet$pred_avg>0.5,'Y','N'))
  • Majority Voting: In majority voting, we’ll assign the prediction for the observation as predicted by the majority of models. Since we have three models for a binary classification task, a tie is not possible.
#The majority vote
testSet$pred_majority<-as.factor(ifelse(testSet$pred_rf=='Y' & testSet$pred_knn=='Y','Y',ifelse(testSet$pred_rf=='Y' & testSet$pred_lr=='Y','Y',ifelse(testSet$pred_knn=='Y' & testSet$pred_lr=='Y','Y','N'))))
  • Weighted Average: Instead of taking simple average, we can take weighted average. Generally, the weights of predictions are high for more accurate models. Let’s assign 0.5 to logistic regression and 0.25 to KNN and random forest each.
#Taking weighted average of predictions
testSet$pred_weighted_avg<-(testSet$pred_rf_prob$Y*0.25)+(testSet$pred_knn_prob$Y*0.25)+(testSet$pred_lr_prob$Y*0.5)

#Splitting into binary classes at 0.5
testSet$pred_weighted_avg<-as.factor(ifelse(testSet$pred_weighted_avg>0.5,'Y','N'))

Before proceeding further, I would like you to recall about two important criteria that we previously discussed on individual model accuracy and inter-model prediction correlation which must be fulfilled. In the above ensembles, I have skipped checking for the correlation between the predictions of the three models. I have randomly chosen these three models for a demonstration of the concepts. If the predictions are highly correlated, then using these three might not give better results than individual models. But you got the point. Right?

So far, we have used simple formulas at the top layer. Instead, we can use another machine learning model which is essentially what stacking is. We can use linear regression for making a linear formula for making the predictions in regression problem for mapping bottom layer model predictions to the outcome or logistic regression similarly in case of classification problem.

Moreover, we don’t need to restrict ourselves here, we can also use more complex models like GBM, neural nets to develop a non-linear mapping from the predictions of bottom layer models to the outcome.

On the same example let’s try applying logistic regression and GBM as top layer models. Remember, the following steps that we’ll take:

  1. Train the individual base layer models on training data.
  2. Predict using each base layer model for training data and test data.
  3. Now train the top layer model again on the predictions of the bottom layer models that has been made on the training data.
  4. Finally, predict using the top layer model with the predictions of bottom layer models that has been made for testing data.

One extremely important thing to note in step 2 is that you should always make out of bag predictions for the training data, otherwise the importance of the base layer models will only be a function of how well a base layer model can recall the training data.

Even, most of the steps have been already done previously, but I’ll walk you through the steps one by one again.

 

  • Step 1: Train the individual base layer models on training data

#Defining the training control
fitControl <- trainControl(
method = "cv",
number = 10,
savePredictions = 'final', # To save out of fold predictions for best parameter combinantions
classProbs = T # To save the class probabilities of the out of fold predictions
)

#Defining the predictors and outcome
predictors<-c("Credit_History", "LoanAmount", "Loan_Amount_Term", "ApplicantIncome",
"CoapplicantIncome")
outcomeName<-'Loan_Status'

#Training the random forest model
model_rf<-train(trainSet[,predictors],trainSet[,outcomeName],method='rf',trControl=fitControl,tuneLength=3

#Training the knn model
model_knn<-train(trainSet[,predictors],trainSet[,outcomeName],method='knn',trControl=fitControl,tuneLength=3)

#Training the logistic regression model
model_lr<-train(trainSet[,predictors],trainSet[,outcomeName],method='glm',trControl=fitControl,tuneLength=3)

 

  • Step 2: Predict using each base layer model for training data and test data

#Predicting the out of fold prediction probabilities for training data
trainSet$OOF_pred_rf<-model_rf$pred$Y[order(model_rf$pred$rowIndex)]
trainSet$OOF_pred_knn<-model_knn$pred$Y[order(model_knn$pred$rowIndex)]
trainSet$OOF_pred_lr<-model_lr$pred$Y[order(model_lr$pred$rowIndex)]

#Predicting probabilities for the test data
testSet$OOF_pred_rf<-predict(model_rf,testSet[predictors],type='prob')$Y
testSet$OOF_pred_knn<-predict(model_knn,testSet[predictors],type='prob')$Y
testSet$OOF_pred_lr<-predict(model_lr,testSet[predictors],type='prob')$Y

 

  • Step 3: Now train the top layer model again on the predictions of the bottom layer models that has been made on the training data

First, let’s start with the GBM model as the top layer model.

 

#Predictors for top layer models 
predictors_top<-c('OOF_pred_rf','OOF_pred_knn','OOF_pred_lr') 

#GBM as top layer model 
model_gbm<- 
train(trainSet[,predictors_top],trainSet[,outcomeName],method='gbm',trControl=fitControl,tuneLength=3)

Similarly, we can create an ensemble with logistic regression as the top layer model as well.

#Logistic regression as top layer model
model_glm<-
train(trainSet[,predictors_top],trainSet[,outcomeName],method='glm',trControl=fitControl,tuneLength=3)

 

  • Step 4: Finally, predict using the top layer model with the predictions of bottom layer models that has been made for testing data

#predict using GBM top layer model
testSet$gbm_stacked<-predict(model_gbm,testSet[,predictors_top])

#predict using logictic regression top layer model
testSet$glm_stacked<-predict(model_glm,testSet[,predictors_top])

Great! You made your first ensemble.

Note it’s really important to choose the models for the ensemble wisely to get the best out of the ensemble. The two thumb rules that we discussed will greatly help you in that.

 

5. Additional resources

By now, you might have developed an in-depth conceptual as well as practical knowledge of ensembling. I would like to encourage you to practice this on machine learning hackathons on Analytics Vidhya, which you can find here.

You’ll probably find this article on top five questions related to ensembling helpful.

Also, if you missed out on the skilltest on ensembling, you can check your understanding of ensembling concepts here.

 

End Notes

Ensembling is a very popular and effective technique that is very frequently used by data scientists for beating the accuracy benchmark of even the best of individual algorithms. More often than not it’s the winning recipe in hackathons. The more you’ll use ensembling, the more you’ll admire its beauty.

Did you enjoy reading this article?  Do share your views in the comment section below. If you have any doubts / questions feel free to drop them in the comments below.

Learn, compete, hack and get hired!

sauravkaushik8 25 Jun, 2019

Frequently Asked Questions

Lorem ipsum dolor sit amet, consectetur adipiscing elit,

Responses From Readers

Clear

Manish Saraswat
Manish Saraswat 15 Feb, 2017

Awesome tutorial. very nicely explained!

Ankit Gupta
Ankit Gupta 15 Feb, 2017

Nicely written!

Albert
Albert 15 Feb, 2017

Thank you for the great article! Although you could have shared the result of your ensembles...

Kumar
Kumar 15 Feb, 2017

Simply written and easy to understand,.

Ravi Chawla
Ravi Chawla 15 Feb, 2017

Kudos, Very well written.

Gonzalo Moreno
Gonzalo Moreno 15 Feb, 2017

thank you. Very good. However I would like to check the accuracy of the ensembled model and you did not show it.

Yogesh Srivastava
Yogesh Srivastava 15 Feb, 2017

I really wanted to compliment in being so transparent in sharing this knowledge and learning platform. You are strengthening the often overlooked notion when in reality sharing knowledge makes us better. I am an avid reader of your posts as it helps me understand the changing world of advanced analytics and have informed discussions with data scientists. Thanks again Yogi

Mathan
Mathan 16 Feb, 2017

Simple & Fantastic!

Sanjay
Sanjay 16 Feb, 2017

Great article!!It would be great if you could also explain about Monte-Carlo simulation for finding best ensemble weights.

Amit Gupta
Amit Gupta 17 Feb, 2017

Thanks Saurav. This article came at the right time. I was looking for stacking and blending example and read many article on the subject but didn't find complete hands on example on R. So struggling thru but this article initialized me with the basic.That's why I love this site!!!!.

Himanshu Gupta
Himanshu Gupta 17 Feb, 2017

Thanks Saurav, I have been searching for something like this to learn ensemble modeling and glad i found it. Kudos to your work!

Kunal Bhowmick
Kunal Bhowmick 18 Feb, 2017

Very well explained .

Arun Pabians
Arun Pabians 19 Feb, 2017

Hi Saurav, You are nicely presented about ensemble modelling. I really learned lot. I have clarification here.If i look at the disadvantages of ensemble modelling,It giving the sense like no use of ensemble modelling in the real time application an other than data hackthon. Correct me if i am wrong. Regards Arun

Charmi Dholakia
Charmi Dholakia 19 Feb, 2017

what have we done without you!! I liked this article so much ..could easily understand.. Thanks

Mahesh
Mahesh 22 Feb, 2017

This is Just awesome explanation.Liked Analytics Vidhya a lot.

Akash Ansari
Akash Ansari 27 Feb, 2017

Thanks Saurav for keeping it simple. Well, I find that KNN, in terms of accuracy, is the best among others including the stacked models. And glm_stacked is exactly same as the logistic regression.

Jain
Jain 28 Feb, 2017

Hi , Thanks for the article :) When i am trying to use : testSet$pred_rf_prob<-predict(object = model_rf,testSet[,predictors],type='prob') I am getting zero probabilities. Why it is giving zero probabilities and what it actually means?

Rohit Joshi
Rohit Joshi 02 Mar, 2017

For the part : model_rf<-train(trainset[, predictors],trainset[,outcomeName],method='rf',trControl=fitControl,tuneLength=3) I am getting error msg as : Error in trainset[, predictors] : incorrect number of dimensions Why?

Georgios Sarantitis
Georgios Sarantitis 20 Mar, 2017

I am starting to get a Machine Learning overdose here..! As a newbie to Data Science I initially thought that a simple regression model, a decision tree or a knn setting would do the job. And I just found out the reason that I cannot reach the highest positions in the leaderboard by using only them! Now I have to study some more I guess....Anyway, thank you very much for your wonderful presentation and explanation of the topic, you have pointed me the way ahead!

Aakansha
Aakansha 26 Mar, 2017

Very Nicely explained

germmie
germmie 29 May, 2017

Great article! Want to clarify something though regarding the stacking procedure: There are others that does this: they train the base model i.e RF, KNN and logistic with the testset (the 25% of the trainset) and later, fit the predicted values as features for the next layer model i.e GBM. With this model, we fit the actual test data (that is not read in this tutorial) to get the predicted submission. Is this correct?

Sayak Paul
Sayak Paul 17 Jun, 2017

Fantastic tutorial. Explained with such ace. Thank you very much. Things are starting to make sense now. Thank you again. I really liked the way you distinguished the Base models and Top model in the diagram and accordingly did the code. Cheers!

Pouria
Pouria 26 Jun, 2017

Hi Saurav, Thank you so much for sharing you knowledge here. I have learned a bunch from your post. I just wanted to point out that your method for extracting predictions for one particular level won't work out, if we have more than 2 levels. I think the better approach for such cases or overall (even for the 2 level cases), would be to just do the predict(), without declaring the type 'prob'. Below is the sections, I'm referring to (using $Y): #Predicting the out of fold prediction probabilities for training data trainSet$OOF_pred_rf<-model_rf$pred$Y[order(model_rf$pred$rowIndex)] #Predicting probabilities for the test data testSet$OOF_pred_rf<-predict(model_rf,testSet[predictors],type='prob')$Y Please, correct me if I may be missing something here and you may have better overall approach for such cases (mostly with more than 2 levels). Thank you :)

Vyom Bani
Vyom Bani 05 Jul, 2017

Is there the same article in python? If not, can you add it please! Thank You.

Akhil Jain
Akhil Jain 28 Jul, 2017

Saurav..i would like to thank you for your post..its very simple to understand and equally simple to implement. However, i faced some minor issues: When we predict the probabilities (for taking the average), we actually get 2 columns.. for e.g. CVSet$pred_rf_prob = predict(object = model_rf,CVSet[,input],type='prob') will actually create 2 columns in the data frame. CVSet$pred_rf_prob.N and CVSet$pred_rf_prob.Y which will contain the probability of getting a N or a Y We can then take the average of the 3 models..but now come 2 problems 1. the command CVSet$pred_avg = ifelse(CVSet$pred_avg>0.5,'Y','N') actually does the opposite...it gives N when prob.Y>0.5 and Y when prob.Y<0.5, which leads me to understand that the conputation is happening on the N column and not on the Y column 2. when converting to Y or N, the value is stored in the prob.N column, leaving prob.Y column as empty.....problem now comes when converting to factor CVSet$pred_avg = factor(CVSet$pred_avg) gives an error "replacement has 306 rows, data has 153" Also, can somebody please check: whether as.factor() or factor() should be used..

j.akhil.j
j.akhil.j 29 Jul, 2017

How do you check for the correlation between the predictions of the three models ?? any useful links?

saurabh
saurabh 29 Jul, 2017

In step 2 why did we use order?? Please help #Predicting the out of fold prediction probabilities for training data trainSet$OOF_pred_rf<-model_rf$pred$Y[order(model_rf$pred$rowIndex)] trainSet$OOF_pred_knn<-model_knn$pred$Y[order(model_knn$pred$rowIndex)] trainSet$OOF_pred_lr<-model_lr$pred$Y[order(model_lr$pred$rowIndex)]

Sam
Sam 29 Aug, 2017

Hi Saurav, Thanks for such valuable post, any idea how to do it using java? Regards

K
K 18 Dec, 2017

Hi, In step 3 #Predictors for top layer models predictors_top<-c('OOF_pred_rf','OOF_pred_knn','OOF_pred_lr') #GBM as top layer model model_gbm<- train(trainSet[,predictors_top],trainSet[,outcomeName],method='gbm',trControl=fitControl,tuneLength=3) when i try to use the code on different data set , i get a error Error in `[.data.frame`(trainSet, , predictors_top) : undefined columns selected Have you defined ('OOF_pred_rf','OOF_pred_knn','OOF_pred_lr') ? how can we use them in model_gbm. They dont consist of any data. Please clarify. Thanks

Shubham Soni
Shubham Soni 28 Dec, 2017

Hi Saurav, Thanks for the wonderful article, little correction (if I understood ensemble correctly) in the last step--> Step 4: Finally, predict using the top layer model with the predictions of bottom layer models that has been made for testing data I think the last line should be using "gbm_stacked" data for glm prediction As follows: #predict using logictic regression top layer model testSet$glm_stacked<-predict(model_glm,testSet[,gbm_stacked]) let me know if I am being wrong here. Thanks Shubham

Prasanta Panja
Prasanta Panja 28 Apr, 2018

Nice explaination!

Rupesh
Rupesh 25 May, 2018

Excellent!! Very well written!!

JUNGHO, BAE
JUNGHO, BAE 02 Mar, 2022

Excellent!! Very well explained. Thank tou!!! I have the following error. Please review. Please Help me. ... > confusionMatrix(testSet$Loan_Status,testSet$pred_rf) Error: `data` and `reference` should be factors with the same levels.