Entropy & Information Gain
At each split, the tree asks: "Which question reduces diagnostic uncertainty the most?"
Entropy measures impurity — how mixed malignant/benign cases are at a node:
H = −Σ pᵢ log₂(pᵢ)
H = 0 → pure (all one class)
H = 1 → maximally mixed (50/50)
Information Gain is the entropy drop after a split:
IG = H(parent) − Σ (|child|/|parent|)·H(child)
The tree always picks the split with the
highest IG — the most informative question.