Entropy is the average level of โsurpriseโ or โuncertaintyโ in a probability distribution. Intuitively, the surprise of an individual event
Note that we apply the log to change the bound from
By combining all events together with an expectation, we get the entropy equation
Another way to interpret this is the expected number of bits needed to encode
Info
Uncertainty is a measure of the variance of a distribution. A distribution with high entropy or uncertainty would be roughly uniform.
Below is a graph of the entropy of binary variable
Conditional Entropy
Just like with a standard distribution, entropy can be measured for a conditional distribution
Conditional entropy is the expectation of this entropy for all
More intuitively, we can derive