Skip to main content

WHAT IS BOOSTING

                                                 BOOSTING

What is Boosting?

Boosting and Bagging algorithms are belonging to the ensemble learning techniques.

Boosting refers to converting weak learners into Strong learners.

what are weak learners?

if our base model predicts wrongly(like spam mail as not spam) the features are called weak learners.

weak learners are nothing but the 

To convert Weak learners into strong learners we generally apply weights.

Let's take an example.

Note:-The base learner may be any model like decision Tree, SVM, etc...

we have a dataset with 100 features, in the ensemble technique we have base learners here what we do is we take some amount of data in the first base learner.

in our first base learner model, the model predicts some features are not predicted well These wrongly predicted features are called weak learners.

In the next base learner, we take this weak learner and some other features from the dataset we will do the same procedure here also we will get some other weak learners.

After applying base learners at some stage we get a pure model.

To convert weak learner to strong learner, we’ll combine the prediction of each weak learner using methods like:

> weighted average.

Considering prediction has a higher vote.


in this way, the boosting algorithm convert weak learners into strong learners

THE END


 

Comments

Popular posts from this blog

Loss Functions | MSE | MAE | RMSE

            Performance Metrics The various metrics used to evaluate the results of the prediction are : Mean Squared Error(MSE) Mean Absolute error(MAE) Root-Mean-Squared-Error(RMSE) Adjusted R² Mean Squared Error: Mean Squared error is one of the most used metrics for regression tasks. MSE is simply the average of the squared difference between the target value and value predicted by the regression model.  As it squares the differences and  penalizes (punish)even a small error which leads to over-estimation of how bad the model is. It is preferred more than other metrics because it is differentiable and hence can be optimized better. in the above formulae, y=actual value and ( yhat) means predicted value by the model. RMSE(Root Mean Squared Error: This is the same as MSE (Mean Squared Error) but the root of the value is considered while determining the accuracy of the model. It is preferred more in some cases because the errors are first...

LSTM

                                                                                  LSTM  we will discuss now each and every stage with the help of the above diagram. State1:Memory View The memory view is responsible for remembering and forget the information based on the context of an input. (you didn't get it, wait now you will understand). In the above diagram, the memory view is the top line.   key points( Ct-1, X, +,  and Ct) . The input is an old memory, X is multiplication which forgets the useless information from the old memory, and " +"   addition lets merge all these things. when we multiply the old memory with '0' the old memory will "0" or if we multiply with vector "1" The old memory won't change. ( what ...

KNN Interview Questions

                           KNN interview questions 1) Which of the following distance metric can not be used in k-NN? A) Euclidean Distance B) Manhatten Distance c) Hamming Distance E) Minkowski Distance F) Jaccard Distance G) All the above Answer:- G All of these distance metric can be used as a distance metric for KNN 2)Knn is for regression or classification? Answer:- Knn is used for both classification and regression problems. 3) When we use Manhatten Distance? Answer:-Manhatten distance is used for continuous variables. 4) You have given the following 2 statements, find which of these options is/are true in case of k-NN? In the case of very large value of k , we may include points from other classes into the neighborhood, so it leads to overfitting. In case of too small value of k the algorithm is very sensitive to noise.(it will affect our model performance). Answer:-The above two points are answers. 5...