Ear Shape Split
Root Variance: 20.51 Weighted Variance: 11.67 Variance Reduction: 8.84
- Best reduction
Decision trees can be generalized beyond classification to handle regression tasks where the goal is predicting numerical values rather than discrete categories.
Previous: Predict if animal is cat (classification) Now: Predict weight of animal (regression)
Input Features (unchanged):
Output Variable:
Predicting a number rather than a category.
Sample regression tree:
Note: Same feature can appear in multiple branches - this is perfectly valid in decision trees.
Classification vs. Regression Difference:
Leaf Node Calculation: Average of training examples reaching that node
Example Leaf Node:
Additional Examples:
Classification Trees: Minimize entropy (measure of class impurity) Regression Trees: Minimize variance (measure of numerical spread)
Variance Definition: How widely a set of numbers varies
Example Comparisons:
Interpretation: Higher variance indicates more spread in values, suggesting need for further splitting.
Similar structure to classification, but using variance instead of entropy:
Split Evaluation:
Weighted Variance = w^left × Variance(left) + w^right × Variance(right)
Example: Ear shape split
Formula:
Variance Reduction = Root Variance - Weighted Variance After Split
Calculation Examples:
Ear Shape Split
Root Variance: 20.51 Weighted Variance: 11.67 Variance Reduction: 8.84
Face Shape Split
Root Variance: 20.51 Weighted Variance: 19.87 Variance Reduction: 0.64
Whiskers Split
Root Variance: 20.51 Weighted Variance: 14.29 Variance Reduction: 6.22
Selection Rule: Choose feature with largest variance reduction
Example Result: Ear shape (8.84) > Whiskers (6.22) > Face shape (0.64) Decision: Split on ear shape
After selecting ear shape:
Similar to classification trees:
Aspect | Classification Trees | Regression Trees |
---|---|---|
Output | Categories/Classes | Numerical Values |
Leaf Prediction | Most common class | Average of values |
Splitting Criterion | Entropy/Information Gain | Variance Reduction |
Impurity Measure | Class mixture | Value spread |
Regression trees extend the power of decision trees beyond classification, enabling prediction of continuous numerical outcomes while maintaining the interpretable tree structure and systematic splitting approach.