Code Monkey home page Code Monkey logo

ml-ensembles-methods's Introduction

ML-Ensembles-Methods

The goal of Ensemble methods is to combine the predictions of several base estimators built with a given learning algorithm in order to improve generalizability / robustness over a single estimator.OR Ensemble methods is a machine learning technique that combines several base models in order to produce one optimal predictive OR Ensemble is the art of combining diverse set of learners (individual models) together to improvise on the stability and predictive power of the model

Two families of ensemble methods are usually distinguished:

  1. averaging methods, The driving principle is to build several estimators independently and then to average their predictions. On average, the combined estimator is usually better than any of the single base estimator because its variance is reduced. Examples: Bagging methods, Forests of randomized trees, โ€ฆ

  2. boosting methods, Base estimators are built sequentially and one tries to reduce the bias of the combined estimator. The motivation is to combine several weak models to produce a powerful ensemble. Examples: AdaBoost, Gradient Tree Boosting

The three most popular methods for combining the predictions from different models are:

  1. Bagging - Building multiple models (typically of the same type) from different subsamples of the training dataset.

  2. Boosting - Building multiple models (typically of the same type) each of which learns to fix the prediction errors of a prior model in the chain.

  3. Voting - Building multiple models (typically of differing types) and simple statistics (like calculating the mean) are used to combine predictions.

  4. Bagging Algorithms Bootstrap Aggregation or bagging involves taking multiple samples from your training dataset (with replacement) and training a model for each sample. The final output prediction is averaged across the predictions of all of the sub-models.

The three bagging models covered in this section are as follows:

(a). Bagged Decision Trees Bagging performs best with algorithms that have high variance. A popular example are decision trees, often constructed without pruning

(b). Random Forest Samples of the training dataset are taken with replacement, but the trees are constructed in a way that reduces the correlation between individual classifiers. Specifically, rather than greedily choosing the best split point in the construction of the tree, only a random subset of features are considered for each split.

(c).Extra Trees Extra Trees are another modification of bagging where random trees are constructed from samples of the training dataset.

  1. Boosting Algorithm Boosting ensemble algorithms creates a sequence of models that attempt to correct the mistakes of the models before them in the sequence.Once created, the models make predictions which may be weighted by their demonstrated accuracy and the results are combined to create a final output prediction.
The two most common boosting ensemble machine learning algorithms are:

(a). AdaBoost AdaBoost was perhaps the first successful boosting ensemble algorithm. It generally works by weighting instances in the dataset by how easy or difficult they are to classify, allowing the algorithm to pay more or less attention to them in the construction of subsequent models.

(b). Stochastic Gradient Boosting(the best) Stochastic Gradient Boosting (also called Gradient Boosting Machines) are one of the most sophisticated ensemble techniques. It is also a technique that is proving to be perhaps of the the best techniques available for improving performance via ensembles

RESOURCES: 1.https://www.analyticsvidhya.com/blog/2015/08/introduction-ensemble-learning/

ml-ensembles-methods's People

Contributors

davisy avatar

Watchers

James Cloos avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.