Processing math: 100%

MathJax

Friday, February 13, 2015

Entropy maximization

One interesting and well known result is that out of all finite alphabet distributions, the uniform distribution achieves maximum entropy.

The Laplace's principle of insufficient reasoning, calls for assuming uniformity unless there is additional information.

Entropy maximization is the equivalent of minimizing the KL-Divergence between the distribution p and the uniform distribution.

More precisely, let X{α1,,αn} be a random variable with finite alphabet, given a family of distribution P and the uniform distribution u,
arg minpPDKL(pu)=arg maxpPH(p) where H(p) is the entropy.

Proof:
D(pu)=ipilogpi+(ipi)log(n)=log(n)H(p)
Note
This is true in general for random variables with finite support.  For RVs with infinite support,  additional constraint is required.  See Gibbs measure...

No comments: