Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Weak Classifier

I am trying to implement an application that uses AdaBoost algorithm. I know that AdaBoost uses set of weak classifiers, but I don't know what these weak classifiers are. Can you explain it to me with an example and tell me if I have to create my own weak classifiers or I'm suppoused to use some kind of algorithm?

like image 882
AjMeen Avatar asked Aug 23 '12 17:08

AjMeen


People also ask

What is a weak classifier?

A weak classifier is a model for binary classification that performs slightly better than random guessing. A weak learner produces a classifier which is only slightly more accurate than random classification.

What are strong and weak classifiers?

Weak classifiers (or weak learners) are classifiers which perform only slightly better than a random classifier. These are thus classifiers which have some clue on how to predict the right labels, but not as much as strong classifiers have like, e.g., Naive Bayes, Neurel Networks or SVM.

What is a weak classifier a classifier with an error rate?

Weak learners are algorithms whose error rate is slightly under 50% as illustrated below : A weak classifier, achieving just under 50% error rate.

What are weak learners in ML?

The term Weak Learner refers to simple models that do only slightly better than random chance. Boosting algorithms start with a single weak learner (tree methods are overwhelmingly used here), but technically, any model will do.


2 Answers

Weak classifiers (or weak learners) are classifiers which perform only slightly better than a random classifier. These are thus classifiers which have some clue on how to predict the right labels, but not as much as strong classifiers have like, e.g., Naive Bayes, Neurel Networks or SVM.

One of the simplest weak classifiers is the Decision Stump, which is a one-level Decision Tree. It selects a threshold for one feature and splits the data on that threshold. AdaBoost will then train an army of these Decision Stumps which each focus on one part of the characteristics of the data.

like image 134
Sicco Avatar answered Nov 27 '22 09:11

Sicco


When I used AdaBoost, my weak classifiers were basically thresholds for each data attribute. Those thresholds need to have a performance of more than 50%, if not it would be totally random.

Here is a good presentation about Adaboost and how to calculate those weak classifiers: https://user.ceng.metu.edu.tr/~tcan/ceng734_f1112/Schedule/adaboost.pdf

like image 39
marc_ferna Avatar answered Nov 27 '22 10:11

marc_ferna