SKETCHINESS ALERT

1. Decision Tree

  1. Recursive partitioning

  2. Pruning & Regularization

    When you split too much → overfitting

    When you split too few → underfitting

    Therefore, we need to find the moderate level between the two

    $$ Cost(T)=Err(T)-\alpha L(T) $$

2. Random Forest

  1. Intuition & main idea:

    Decision tree model only builds a single tree. This has a good interpretability but is prone to overfitting → What if we do subsampling, train multiple trees, and compute the mean (or voting)?

  2. Implementation

    For each tree:

3. Adaboost

Serial boosting: Put more weight on wrongly classified instances.

4. Gradient Boosting

  1. Intuition:

    Recall that linear regression regards the residual as something predictable. However, GBMs treat residuals as something “learnable”

    $$ \hat{y}=f_1(x) \newline f_2(x)=y-f_1(x) \newline f_3(x)=y-f_1(x)-f_2(x) \newline ... $$