Processing math: 100%

MathJax

Thursday, February 12, 2015

A Side Path to Statistical Mechanics

Consider a physical system with many degrees of freedom, that can reside in any one of a large number of possible states.  Let pi denote the probability of occurrence of state i, for example, with the following properties:
pi0for alliand 
ipi=1 Let Ei denote the energy of the system when it is in state i.  A fundamental results from statistical mechanics tells us that when the system is in thermal equilibrium with its surrounding environment, statie i occurs with a probability define by 
pi=1Zexp(EikBT) where T is the absolute temperature in kelvins, kB is the Boltzmann's constant, and Z is a constant that is independent of all states.  The partition function Z is the normalizing constant with 
Z=iexp(EikBT).  The probability distribution is called the Gibbs distribution.

Two interesting properties of the Gibbs distribution are:
  1. States of low energy have a higher probability of occurrence than states of high energy.
  2. As the temperature T is reduced, the probability is concentrated on a smaller subset of low-energy states.
In the context of neural networks, the parameter T may be viewed as a pseudo-temperature that controls thermal fluctualtions representing the effect of "synaptic noise" in a neuron.  Its precise scale is irrelevant.  We can redefine the probability pi and partition function Z as
pi=1Zexp(EiT) and
Z=iexp(EiT) where T is referred to simply as the temperature of the system. 

Note that logpi may be viewed as a form of "energy" measured at unit temperature.

Free Energy and Entropy

The Helmholtz free energy of a physical system, denoted by F, is defined in terms of the partition function Z as follows
F=TlogZ.  The average energy of the system is defined by
<E>=ipiEi The difference between the average energy and free energy is
<E>F=Tipilogpi which we can rewrite in terms of entropy H
<E>F=TH or, equivalently,
F=<E>TH.  The entropy of any systems tend to increase until it reaches an equilibrium, and therefore the free energy of the system will reach a minimum.

This is an important principle called the principle of minimal free energy.

No comments: