pi≥0for alliand
∑ipi=1 Let Ei denote the energy of the system when it is in state i. A fundamental results from statistical mechanics tells us that when the system is in thermal equilibrium with its surrounding environment, statie i occurs with a probability define by
pi=1Zexp(−EikBT) where T is the absolute temperature in kelvins, kB is the Boltzmann's constant, and Z is a constant that is independent of all states. The partition function Z is the normalizing constant with
Z=∑iexp(−EikBT). The probability distribution is called the Gibbs distribution.
Two interesting properties of the Gibbs distribution are:
- States of low energy have a higher probability of occurrence than states of high energy.
- As the temperature T is reduced, the probability is concentrated on a smaller subset of low-energy states.
In the context of neural networks, the parameter T may be viewed as a pseudo-temperature that controls thermal fluctualtions representing the effect of "synaptic noise" in a neuron. Its precise scale is irrelevant. We can redefine the probability pi and partition function Z as
pi=1Zexp(−EiT) and
Z=∑iexp(−EiT) where T is referred to simply as the temperature of the system.
Note that −logpi may be viewed as a form of "energy" measured at unit temperature.
The Helmholtz free energy of a physical system, denoted by F, is defined in terms of the partition function Z as follows
F=−TlogZ. The average energy of the system is defined by
<E>=∑ipiEi The difference between the average energy and free energy is
<E>−F=−T∑ipilogpi which we can rewrite in terms of entropy H
<E>−F=TH or, equivalently,
F=<E>−TH. The entropy of any systems tend to increase until it reaches an equilibrium, and therefore the free energy of the system will reach a minimum.
This is an important principle called the principle of minimal free energy.
No comments:
Post a Comment