首页 > 其他 > 详细

机器学习笔记(Washington University)- Classification Specialization-week five

时间:2017-05-17 22:49:44      阅读:330      评论:0      收藏:0      [点我收藏+]

1. Ensemble classifier 

Each classifier votes on prediction

Ensemble model = sign(w1f1(xi) + w2f2(xi) + w3f3(xi))

www3 is the learning coefficients

f1(xi), f2(xi), f3(xi)) is three classifiers

 

2. Boosting

Focus on hard or more important pointsand keep adding new classfier.

技术分享

Boosting is more robust to overfitting but we still need carefully to choose boosting captical T

using validation set or cross validation.

 

3. Adaboost

1. Start with weight for all points: αi = 1/N

For t = 1 ... T

  • Learn ft(x) with data weights αi
  • Compute coefficient w
    • Note :

      Adaboost use the formual below to compute coefficient wt of classifier ft(x)

      wt  = 1/2*ln(1- weighted_error(ft)/weighted_error(ft))

  • Recompute weights αi
    •   α= αie-Wt, if ft(xi)=yi else αieWt
  • Normalizing weights:
    •   αi = αi / (α1 +α2 ...  αN)

    

Final model predicts the value by:

y = sign(wf1(x) + wft(x) ... wfT(x))

 

Weighted classification error:

weighted_error = total weight of mistakes / total weights of all data points

 

Normalizing weights αi

normalize weights to add up to 1 after every iterationn

α= αi / (α1 +α2 ...  αN)

 

4. Adaboost Theorem

if we can find a weak leatner with weighted_error < 0.5 (beat random guess) at every iteration t,

 the training error of boosted classifier goes to zero as the iterations of boosting goes to infinity.

 

机器学习笔记(Washington University)- Classification Specialization-week five

原文:http://www.cnblogs.com/climberclimb/p/6864549.html

(0)
(0)
   
举报
评论 一句话评论(0
关于我们 - 联系我们 - 留言反馈 - 联系我们:wmxa8@hotmail.com
© 2014 bubuko.com 版权所有
打开技术之扣,分享程序人生!