The Laplace's principle of insufficient reasoning, calls for assuming uniformity unless there is additional information.
Entropy maximization is the equivalent of minimizing the KL-Divergence between the distribution p and the uniform distribution.
More precisely, let X∈{α1,⋯,αn} be a random variable with finite alphabet, given a family of distribution P and the uniform distribution u,
arg minp∈PDKL(p‖u)=arg maxp∈PH(p) where H(p) is the entropy.
Proof:
D(p‖u)=∑ipilogpi+(∑ipi)log(n)=log(n)−H(p)
Note:
This is true in general for random variables with finite support. For RVs with infinite support, additional constraint is required. See Gibbs measure...
No comments:
Post a Comment