Example Selection Measure:
Information Entropy
 
  • Let p(Cij) = Prob(Cj|A=ai) = Cij/Si
  • Entropy of class distribution over partition:

  • Attribute A with values { a1, a2, ..., ar }

    Entropy(Si) max when all Cij are equal;
    Entropy(Si) min when for some i,j Cij=Si