- Statistical mechanics
**Statistical mechanics**is the application ofprobability theory , which includes mathematical tools for dealing with large populations, to the field ofmechanics , which is concerned with the motion of particles or objects when subjected to a force.Statistical mechanics, sometimes calledstatistical physics , can be viewed as a subfield ofphysics andchemistry . Pioneers in establishing the field wereLudwig Boltzmann andJosiah Willard Gibbs .It provides a framework for relating the microscopic properties of individual atoms and molecules to the macroscopic or bulk properties of materials that can be observed in everyday life, therefore explaining

thermodynamics as a natural result of statistics and mechanics (classical and quantum) at the microscopic level. In particular, it can be used to calculate the thermodynamic properties of bulk materials from the spectroscopic data of individual molecules.This ability to make macroscopic predictions based on microscopic properties is the main advantage of statistical mechanics over

thermodynamics . Both theories are governed by the second law of thermodynamics through the medium ofentropy . However,entropy inthermodynamics can only be known empirically, whereas in statistical mechanics, it is a function of the distribution of the system on its micro-states.**Fundamental postulate**The fundamental postulate in statistical mechanics (also known as the "equal a priori probability postulate") is the following:

:"Given an isolated system in equilibrium, it is found with equal probability in each of its accessible microstates."

This postulate is a fundamental assumption in statistical mechanics - it states that a system in equilibrium does not have any preference for any of its available microstates. Given Ω microstates at a particular energy, the probability of finding the system in a particular microstate is "p" = 1/Ω.

This postulate is necessary because it allows one to conclude that for a system at equilibrium, the thermodynamic state (macrostate) which could result from the largest number of microstates is also the most probable macrostate of the system.

The postulate is justified in part, for classical systems, by

Liouville's theorem (Hamiltonian) , which shows that if the distribution of system points through accessiblephase space is uniform at some time, it remains so at later times.Similar justification for a discrete system is provided by the mechanism of

detailed balance .This allows for the definition of the "information function" (in the context of

information theory ):...:$I\; =\; -\; sum\_i\; ho\_i\; ln\; ho\_i\; =\; langle\; ln\; ho\; angle.$

When all rhos are equal, I is minimal, which reflects the fact that we have minimal information about the system. When our information is maximal, i.e. one rho is equal to one and the rest to zero (we know what state the system is in), the function is maximal." [Comments: It should be "When all rhos are equal, I is maximal...", please have a check] "This "information function" is the same as the

**reduced entropic function**in thermodynamics.**Microcanonical ensemble**In microcanonical ensemble N, V and E are fixed. Since the

second law of thermodynamics applies toisolated systems, the first case investigated will correspond to this case. The "Microcanonical ensemble" describes anisolated system.The

entropy of such a system can only increase, so that the maximum of itsentropy corresponds to an equilibrium state for the system.Because an

isolated system keeps a constant energy, the totalenergy of the system does not fluctuate. Thus, the system can access only those of its micro-states that correspond to a given value "E" of the energy. Theinternal energy of the system is then strictly equal to itsenergy .Let us call $Omega(E)$ the number of micro-states corresponding to this value of the system's energy. The macroscopic state of maximal

entropy for the system is the one in which all micro-states are equally likely to occur, with probability $\{1/\{Omega\; (E)$, during the system's fluctuations.::$S=-k\_Bsum\_\{i=1\}^\{Omega\; (E)\}\; left\; \{\; \{1over\{Omega\; (E)\; ln\{1over\{Omega\; (E)\; ight\; \}\; =k\_Bln\; left(Omega\; (E)\; ight)$:where :$S$ is the system

entropy ,:$k\_B$ isBoltzmann's constant **Canonical ensemble**In canonical ensemble N, V and T are fixed. Invoking the concept of the canonical ensemble, it is possible to derive the probability $P\_i$ that a macroscopic system in

thermal equilibrium with its environment, will be in a given microstate with energy $E\_i$ according to theBoltzmann distribution :::$P\_i\; =\; \{e^\{-eta\; E\_i\}over\{sum\_j^\{j\_\{maxe^\{-eta\; E\_j\}$

:where $eta=\{1over\{kT$,

The temperature $T$ arises from the fact that the system is in thermal equilibrium with its environment. The probabilities of the various microstates must add to one, and the

normalization factor in the denominator is the canonical partition function:: $Z\; =\; sum\_j^\{j\_\{max\; e^\{-eta\; E\_j\}$

where $E\_i$ is the energy of the $i$th microstate of the system. The partition function is a measure of the number of states accessible to the system at a given temperature. The article

canonical ensemble contains a derivation of Boltzmann's factor and the form of the partition function from first principles.To sum up, the probability of finding a system at temperature $T$ in a particular state with energy $E\_i$ is

: $P\_i\; =\; frac\{e^\{-eta\; E\_i\{Z\}$

**Thermodynamic Connection**The partition function can be used to find the expected (average) value of any microscopic property of the system, which can then be related to macroscopic variables. For instance, the expected value of the microscopic energy $E$ is "interpreted" as the microscopic definition of the thermodynamic variable internal energy $U$., and can be obtained by taking the derivative of the partition function with respect to the temperature. Indeed,

: $langle\; E\; angle=\{sum\_i\; E\_i\; e^\{-eta\; E\_i\}over\; Z\}=-\{1\; over\; Z\}\; \{dZ\; over\; deta\}$

implies, together with the interpretation of $langle\; E\; angle$ as $U$, the following microscopic definition of

internal energy :: $Ucolon\; =\; -\{dln\; Zover\; d\; eta\}.$

The entropy can be calculated by (see

Shannon entropy ): $\{Sover\; k\}\; =\; -\; sum\_i\; p\_i\; ln\; p\_i\; =\; sum\_i\; \{e^\{-eta\; E\_i\}over\; Z\}(eta\; E\_i+ln\; Z)\; =\; ln\; Z\; +\; eta\; U$

which implies that

: $-frac\{ln(Z)\}\{eta\}\; =\; U\; -\; TS\; =\; F$

is the free energy of the system or in other words,

:$Z=e^\{-eta\; F\},$

Having microscopic expressions for the basic thermodynamic potentials $U$ (

internal energy ), $S$ (entropy ) and $F$ (free energy) is sufficient to derive expressions for other thermodynamic quantities. The basic strategy is as follows. There may be an intensive or extensive quantity that enters explicitly in the expression for the microscopic energy $E\_i$, for instance magnetic field (intensive) or volume (extensive). Then, the conjugate thermodynamic variables are derivatives of the internal energy. The macroscopic magnetization (extensive) is the derivative of $U$ with respect to the (intensive) magnetic field, and the pressure (intensive) is the derivative of $U$ with respect to volume (extensive).The treatment in this section assumes no exchange of matter (i.e. fixed mass and fixed particle numbers). However, the volume of the system is variable which means the density is also variable.

This probability can be used to find the average value, which corresponds to the macroscopic value, of any property, $J$, that depends on the energetic state of the system by using the formula:

: $langle\; J\; angle\; =\; sum\_i\; p\_i\; J\_i\; =\; sum\_i\; J\_i\; frac\{e^\{-eta\; E\_i\{Z\}$

where $langle\; J\; angle$ is the average value of property $J$. This equation can be applied to the internal energy, $U$:

: $U\; =\; sum\_i\; E\_i\; frac\{e^\{-eta\; E\_i\{Z\}$

Subsequently, these equations can be combined with known thermodynamic relationships between $U$ and $V$ to arrive at an expression for pressure in terms of only temperature, volume and the partition function. Similar relationships in terms of the partition function can be derived for other thermodynamic properties as shown in the following table; see also the detailed explanation in [

*http://clesm.mae.ufl.edu/wiki.pub/index.php/Configuration_integral_%28statistical_mechanics%29 configuration integral*] .**References***

*

*

*

*

*List of notable textbooks in statistical mechanics **External links*** [

*http://plato.stanford.edu/entries/statphys-statmech/ Philosophy of Statistical Mechanics*] article by Lawrence Sklar for theStanford Encyclopedia of Philosophy .

* [*http://www.sklogwiki.org/ Sklogwiki - Thermodynamics, statistical mechanics, and the computer simulation of materials.*] SklogWiki is particularly orientated towards liquids and soft condensed matter.

*Wikimedia Foundation.
2010.*

### Look at other dictionaries:

**statistical mechanics**— Statistical Mechanics Статистическая механика (статистическая физика) Раздел теоретической физики, в котором изучаются свойства и поведение макроскопических физических тел, состоящих из очень большого числа атомов, молекул, заряженных частиц… … Толковый англо-русский словарь по нанотехнологии. - М.**statistical mechanics**— Physics, Chem. (used with a sing. v.) the science that deals with average properties of the molecules, atoms, or elementary particles in random motion in a system of many such particles and relates these properties to the thermodynamic and other… … Universalium**statistical mechanics**— statistinė mechanika statusas T sritis fizika atitikmenys: angl. statistical mechanics vok. statistische Mechanik, f rus. статистическая механика, f pranc. mécanique statistique, f … Fizikos terminų žodynas**statistical mechanics**— noun plural but usually singular in construction Date: 1885 a branch of mechanics dealing with the application of the principles of statistics to the mechanics of a system consisting of a large number of parts having motions that differ by small… … New Collegiate Dictionary**statistical mechanics**— noun The branch of physics that uses statistical laws to make theoretical predictions about macroscopic systems of particles. Syn: statistical thermodynamics See Also: Boltzmannian … Wiktionary**statistical mechanics**— n. branch of physics that deals with and examines macroscopic systems by implementing statistical principles to their microscopic components … English contemporary dictionary**statistical mechanics**— plural noun [treated as sing.] the description of physical phenomena in terms of a statistical treatment of the behaviour of large numbers of atoms or molecules, especially as regards the distribution of energy among them … English new terms dictionary**statistical mechanics**— noun the branch of physics that makes theoretical predictions about the behavior of macroscopic systems on the basis of statistical laws governing its component particles • Hypernyms: ↑physics, ↑natural philosophy … Useful english dictionary**statistical mechanics**— /stəˌtɪstɪkəl məˈkænɪks/ (say stuh.tistikuhl muh kaniks) noun the application of statistics to the dynamics of large numbers of particles, as in the kinetic theory of gases which deals with large numbers of gas molecules … Australian English dictionary**Partition function (statistical mechanics)**— For other uses, see Partition function (disambiguation). Partition function describe the statistical properties of a system in thermodynamic equilibrium. It is a function of temperature and other parameters, such as the volume enclosing a gas.… … Wikipedia