differential entropy
Let be a probability space, and let , be a function
. The differential entropy is defined as
(1) |
Differential entropy is the continuous version of the Shannon entropy, . Consider first , the uniform 1-dimensional distribution on . The differential entropy is
(2) |
Next consider probability distributions such as the function
(3) |
the 1-dimensional Gaussian. This pdf has differential entropy
(4) |
For a general -dimensional Gaussian (http://planetmath.org/JointNormalDistribution) with mean vector and covariance matrix , , we have
(5) |
where .