Description: AdaBoost boosting systems dollar fusion algorithm is the most popular one, it plainly systems integration approach is a serial train and final weighted cumulative.
Specific process is: Each training example is given equal weight, and the weights satisfy normalization, after the first classifiers after
Calculating a first classifier weights alpha value for each sample and updates right weight training, and then the second classifier training, the same way .......
0, or until the specified error rate training rounds, wherein the label is the calculation of the final prediction system* alpha weighted and then sign (predicted value).
As can be seen, the training process is serial, and weight training examples is always changing, the right of the wrong sample weight continued to increase, the right to correct sample weight decreasing.
AdaBoost algorithm is an element, as well as other methods of boosting popular systems integration, such as bagging and random forest method.
For
To Search:
File list (Check if you may need any files):
adaboost.py
adaboost.readme