|
|
Line 1: |
Line 1: |
| {{Statistical mechanics|cTopic=Ensembles}}
| | There is nothing to tell about myself I think.<br>Enjoying to be a member of wmflabs.org.<br>I just wish I'm useful at all<br><br>Look at my website [http://bestregistrycleanerfix.com/fix-it-utilities fix it utilities] |
| | |
| In [[mathematical physics]], especially as introduced into [[statistical mechanics]] and [[thermodynamics]] by [[Willard Gibbs|J. Willard Gibbs]] in 1902, an '''ensemble''' (also '''statistical ensemble''') is an idealization consisting of a large number of virtual copies (sometimes infinitely many) of a [[system]], considered all at once, each of which represents a possible state that the real system might be in. In other words, a statistical ensemble is a probability distribution for the state of the system.<ref name="gibbs">{{cite book |last=Gibbs |first=Josiah Willard |authorlink=Josiah Willard Gibbs |title=[[Elementary Principles in Statistical Mechanics]] |year=1902 |publisher=[[Charles Scribner's Sons]] |location=New York}}</ref>
| |
| | |
| A '''thermodynamic ensemble''' is a specific variety of statistical ensemble that, among other properties, is in statistical equilibrium (defined below), and is used to derive the properties of [[thermodynamic system]]s from the laws of classical or quantum mechanics.<ref name="Kittel">{{cite book |last=Kittel |first=Charles |authorlink=Charles Kittel |coauthors=Herbert Kroemer |title=Thermal Physics, Second Edition |publisher=W.H. Freeman and Company |year=1980 |isbn=0-7167-1088-9 |location=San Francisco |pages=31 ff}}</ref><ref name="Landau">{{cite book |title=Statistical Physics |last=Landau |first=L.D. |coauthors=Lifshitz, E.M. |isbn=0-08-023038-5 |authorlink=Lev Landau |year=1980 |publisher=Pergamon Press |pages=9 ff }}</ref>
| |
| | |
| This article treats the notion of ensembles in a mathematically rigorous fashion, although relevant physical aspects will be mentioned.
| |
| | |
| == Physical considerations ==
| |
| | |
| The ensemble formalises the notion that an experimenter repeating an experiment again and again under the same macroscopic conditions, but unable to control the microscopic details, may expect to observe a range of different outcomes.
| |
| | |
| The notional size of ensembles in thermodynamics, statistical mechanics and [[quantum statistical mechanics]] can be very large indeed, to include every possible [[microstate (statistical mechanics)|microscopic state]] the system could be in, consistent with its observed [[macroscopic]] properties. But for important physical cases it can be possible to calculate averages directly over the whole of the thermodynamic ensemble, to obtain explicit formulas for many of the thermodynamic quantities of interest, often in terms of the appropriate [[partition function (mathematics)|partition function]] (see below). Some of these results are presented in the [[statistical mechanics]] article.
| |
| | |
| The concept of an equilibrium or stationary ensemble is crucial to some applications of statistical ensembles. Although a mechanical system certainly evolves over time, the ensemble does not necessarily have to evolve. In fact, the ensemble will not evolve if it equally contains all past and future phases of the system. Such a statistical ensemble, that does not change over time, is called ''stationary'' or said to be in ''statistical equilibrium''.<ref name="gibbs"/>
| |
| | |
| === Note on terminology ===
| |
| | |
| *The word "ensemble" is also used for a smaller set of possibilities [[sample (statistics)|sampled]] from the full set of possible states. For example, a collection of [[random walk|walkers]] in a [[Markov chain Monte Carlo]] iteration is called an ensemble in some literature.
| |
| | |
| *The term "ensemble" is often used in physics and physics-influenced literature. In [[probability theory]] literature the term [[probability space]] is more prevalent.
| |
| | |
| == Principal ensembles of statistical thermodynamics ==
| |
| | |
| The study of thermodynamics is concerned with systems which appear to human perception to be "static" (despite the motion of their internal parts), and which can be described simply by a set of macroscopically observable variables. These systems can be described by statistical ensembles that depend on a few observable parameters, and which are in statistical equilibrium. Gibbs noted that different macroscopic constraints lead to different types of ensembles, with particular statistical characteristics. Three important thermodynamic ensembles were defined by Gibbs:<ref name="gibbs"/>
| |
| | |
| * [[Microcanonical ensemble]] or ''NVE'' ensemble—a statistical ensemble where the total energy of the system and the number of particles in the system are each fixed to particular values; each of the members of the ensemble are required to have the same total energy and particle number. The system must remain totally isolated (unable to exchange energy or particles with its environment) in order to stay in statistical equilibrium.<ref name="gibbs"/>
| |
| | |
| * [[Canonical ensemble]] or ''NVT'' ensemble—a statistical ensemble where the energy is not known exactly but the number of particles is fixed. In place of energy, the [[temperature]] is specified. The canonical ensemble is appropriate for describing a closed system which is in, or has been in, weak thermal contact with a heat bath. In order to be in statistical equilibrium the system must remain totally closed (unable to exchange particles with its environment), and may come into weak thermal contact with other systems that are described by ensembles with the same temperature.<ref name="gibbs"/>
| |
| | |
| * [[Grand canonical ensemble]] or ''µVT'' ensemble—a statistical ensemble where neither the energy nor particle number are fixed. In their place, the temperature and [[chemical potential]] are specified. The grand canonical ensemble is appropriate for describing an open system: one which is in, or has been in, weak contact with a reservoir (thermal contact, chemical contact, radiative contact, electrical contact, etc.). The ensemble remains in statistical equilibrium if the system comes into weak contact with other systems that are described by ensembles with the same temperature and chemical potential.<ref name="gibbs"/>
| |
| | |
| The calculations that can be made over each of these ensembles are explored further in their respective articles.
| |
| Other thermodynamic ensembles can be also defined, corresponding to different physical requirements, for which analogous formulae can often similarly be derived.
| |
| | |
| ==Representations for statistical ensembles in statistical mechanics ==
| |
| | |
| The precise mathematical expression for a statistical ensemble has a distinct form depending on the type of mechanics under consideration (quantum or classical). In each case the ensemble is a probability distribution over microstates, but the notion of a "microstate" is considerably different. In quantum mechanics, a set of [[Hilbert space]] basis vectors can be used as microstates, however the ensemble might also include superpositions of these microstates.
| |
| In classical mechanics, the ensemble is instead written as a probability distribution in [[phase space]]; the microstates are the result of partitioning phase space into equal-sized units, although the size of these units can be chosen somewhat arbitrarily.
| |
| | |
| === Requirements for representations ===
| |
| | |
| Putting aside for the moment the question of how statistical ensembles are generated [[operational definition|operationally]], we should be able to perform the following two operations on ensembles ''A'', ''B'' of the same system:
| |
| | |
| * Test whether ''A'', ''B'' are statistically equivalent.
| |
| | |
| * If ''p'' is a real number such that 0 < ''p'' < 1, then produce a new ensemble by probabilistic sampling from ''A'' with probability ''p'' and from ''B'' with probability ''1 – p''.
| |
| | |
| Under certain conditions therefore, [[equivalence class]]es of statistical ensembles have the structure of a convex set.
| |
| | |
| === Quantum mechanical ===
| |
| | |
| {{main|density matrix}}
| |
| | |
| A statistical ensemble in quantum mechanics is most often represented by a [[density matrix]], denoted by {{math|''ρ̂''}}. The density matrix provides a fully general tool that can incorporate both quantum uncertainties (due to superposition) and classical uncertainties (due to lack of knowledge), in a consistent manner. Any physical observable {{math|''X''}} in quantum mechanics can be written as an operator, {{math|''X̂''}}. The expectation value of this operator is given by a [[trace (linear algebra)|trace]]:
| |
| :<math>\langle X \rangle = \operatorname{Tr}(\hat X \rho).</math>
| |
| This can be used to evaluate averages (operator {{math|''X̂''}}), [[variance]]s (using operator {{math|''X̂''<sup>2</sup>}}), [[covariance]]s (using operator {{math|''X̂Ŷ''}}), etc. The density matrix must always have a trace of 1: {{math|Tr ''ρ̂'' {{=}} 1}} (this essentially is the condition that the probabilities must add up to one).
| |
| | |
| In general, the ensemble evolves over time according to the [[von Neumann equation]].
| |
| | |
| Equilibrium ensembles (those that do not evolve over time, {{math|''dρ̂''/''dt'' {{=}} 0}}) can be written solely as a function of conserved variables. For example, the [[microcanonical ensemble]] and [[canonical ensemble]] are functions strictly of the total energy, which is measured by the total energy operator {{math|''Ĥ''}} (Hamiltonian). The grand canonical ensemble is further a function of the particle number, measured by the total particle number operator {{math|''N̂''}}. Such equilibrium ensembles are a [[diagonal matrix]] in the orthogonal basis of states that simultaneously diagonalize each conserved variable. In [[bra-ket notation]], the density matrix is
| |
| :<math>\hat \rho = \sum_i P_i |\psi_i\rangle \langle \psi_i | </math>
| |
| where the {{math|{{!}}''ψ''<sub>''i''</sub>⟩}}, indexed by {{math|''i''}}, are the complete and orthogonal basis. In other words, these states can be considered to be the microstates in quantum mechanics, and the density matrix encodes a probability distribution over these microstates. (Note that in other bases, the density matrix is not necessarily diagonal.)
| |
| | |
| ===Classical mechanical===
| |
| | |
| [[File:Hamiltonian flow classical.gif|frame|Evolution of an ensemble of [[Hamiltonian mechanics|classical]] systems in [[phase space]] (top). Each system consists of one massive particle in a one-dimensional [[potential well]] (red curve, lower figure). The initially compact ensemble becomes swirled up over time.]]
| |
| | |
| In classical mechanics, an ensemble is represented by a probability density function defined over the system's [[phase space]].<ref name="gibbs"/> While an individual system evolves according to [[Hamilton's equations]], the density function (the ensemble) evolves over time according to [[Liouville's theorem (Hamiltonian)|Liouville's equation]].
| |
| | |
| In a [[Hamiltonian mechanics|mechanical system]] with a defined number of parts, the phase space has {{math|''n''}} [[generalized coordinates]] called {{math|''q''<sub>1</sub>, … ''q''<sub>''n''</sub>}}, and {{math|''n''}} associated [[canonical momentum|canonical momenta]] called {{math|''p''<sub>1</sub>, … ''p''<sub>''n''</sub>}}. The ensemble is then represented by a [[joint probability density function]] {{math|''ρ''(''p''<sub>1</sub>, … ''p''<sub>''n''</sub>, ''q''<sub>1</sub>, … ''q''<sub>''n''</sub>)}}.
| |
| | |
| If the number of parts in the system is allowed to vary among the systems in the ensemble (as in a grand ensemble where the number of particles is a random quantity), then it is a probability distribution over an extended phase space that includes further variables such as particle numbers {{math|''N''<sub>1</sub>}} (first kind of particle), {{math|''N''<sub>2</sub>}} (second kind of particle), and so on up to {{math|''N''<sub>''s''</sub>}} (the last kind of particle; {{math|''s''}} is how many different kinds of particles there are). The ensemble is then represented by a [[joint probability density function]] {{math|''ρ''(''N''<sub>1</sub>, … ''N''<sub>''s''</sub>, ''p''<sub>1</sub>, … ''p''<sub>''n''</sub>, ''q''<sub>1</sub>, … ''q''<sub>''n''</sub>)}}. The number of coordinates {{math|''n''}} varies with the numbers of particles.
| |
| | |
| Any physical quantity {{math|''X''}} can be written as a function of the system's phase. The expectation value of any quantity is given by an integral over the entire phase space of this quantity weighted by {{math|''ρ''}}:
| |
| :<math>\langle X \rangle = \sum_{N_1 = 0}^{\infty} \ldots \sum_{N_s = 0}^{\infty} \int \ldots \int \rho X \, dp_1 \ldots dq_n.</math>
| |
| The condition of probability normalization applies, requiring
| |
| :<math>\sum_{N_1 = 0}^{\infty} \ldots \sum_{N_s = 0}^{\infty} \int \ldots \int \rho \, dp_1 \ldots dq_n = 1.</math>
| |
| | |
| Phase space is a continuous space containing an infinite number of distinct physical states within any small region. In order to connect the probability ''density'' in phase space to a probability ''distribution'' over microstates, it is necessary to somehow partition the phase space into blocks that are distributed represent the different states of the system in a fair way. It turns out that the correct way to do this simply results in equal-sized blocks of canonical phase space, and so a microstate in classical mechanics is an extended region in the phase space of canonical coordinates that has a particular volume.<ref group=note>This equal-volume partitioning is a consequence of [[Liouville's theorem (Hamiltonian)|Liouville's theorem]], i. e., the principle of conservation of extension in canonical phase space for Hamiltonian mechanics. This can also be demonstrated starting with the conception of an ensemble as a multitude of systems. See Gibbs' ''Elementary Principles'', Chapter I.</ref> In particular, the probability density function in phase space, {{math|''ρ''}}, is related to the probability distribution over microstates, {{math|''P''}} by a factor
| |
| :<math>\rho = \frac{1}{h^n C} P,</math>
| |
| where
| |
| * {{math|''h''}} is an arbitrary but predetermined constant with the units of {{math|energy×time}}, setting the extent of the microstate and providing correct dimensions to {{math|''ρ''}}.<ref group=note>(Historical note) Gibbs' original ensemble effectively set {{math|''h'' {{=}} 1 [energy unit]×[time unit]}}, leading to unit-dependence in the values of some thermodynamic quantities like entropy and chemical potential. Since the advent of quantum mechanics, {{math|''h''}} is often taken to be equal to [[Planck's constant]] in order to obtain a semiclassical correspondence with quantum mechanics.</ref>
| |
| * {{math|''C''}} is an overcounting correction factor (see below), generally dependent on the number of particles and similar concerns.
| |
| Since {{math|''h''}} can be chosen arbitrarily, the notional size of a microstate is also arbitrary. Still, the value of {{math|''h''}} influences the offsets of quantities such as entropy and chemical potential, and so it is important to be consistent with the value of {{math|''h''}} when comparing different systems.
| |
| | |
| ==== Correcting overcounting in phase space ====
| |
| | |
| Typically, the phase space contains duplicates of the same physical state in multiple distinct locations. This is a consequence of the way that a physical state is encoded into mathematical coordinates; the simplest choice of coordinate system often allows a state to be encoded in multiple ways. An example of this is a gas of identical particles whose state is written in terms of the particles' individual positions and momenta: when two particles are exchanged, the resulting point in phase space is different, and yet it corresponds to an identical physical state of the system. It is important in statistical mechanics (a theory about physical states) to recognize that the phase space is just a mathematical construction, and to not naively overcount actual physical states when integrating over phase space. Overcounting can cause serious problems:
| |
| * Dependence of derived quantities (such as entropy and chemical potential) on the choice of coordinate system, since one coordinate system might show more or less overcounting than another.<ref group=note>In some cases the overcounting error is benign. An example is the [[Charts on SO(3)|choice of coordinate system used for representing orientations of three dimensional objects]]. A simple encoding is the [[3-sphere]] (e. g., unit [[quaternion]]s) which is a [[double covering group|double cover]]—each physical orientation can be encoded in two ways. If this encoding is used without correcting the overcounting, then the entropy will be higher by {{math|''k'' log 2}} per rotatable object and the chemical potential lower by {{math|''kT'' log 2}}. This does not actually lead to any observable error since it only causes unobservable offsets.</ref>
| |
| * Erroneous conclusions that are inconsistent with physical experience, as in the [[mixing paradox]].<ref name="gibbs"/>
| |
| * Foundational issues in defining the [[chemical potential]] and the [[grand canonical ensemble]].<ref name="gibbs"/>
| |
| It is in general difficult to find a coordinate system that uniquely encodes each physical state. As a result, it is usually necessary to use a coordinate system with multiple copies of each state, and then to recognize and remove the overcounting.
| |
| | |
| A crude way to remove the overcounting would be to manually define a subregion of phase space that includes each physical state only once, and then exclude all other parts of phase space. In a gas, for example, one could include only those phases where the particles' {{math|''x''}} coordinates are sorted in ascending order. While this would solve the problem, the resulting integral over phase space would be tedious to perform due to its unusual boundary shape. (In this case, the factor {{math|''C''}} introduced above would be set to {{math|''C'' {{=}} 1}}, and the integral would be restricted to the selected subregion of phase space.)
| |
| | |
| A simpler way to correct the overcounting is to integrate over all phase space, but to reduce the weight of each phase in order to exactly compensate the overcounting. This is accomplished by the factor {{math|''C''}} introduced above, which is a whole number that represents how many ways a physical state can be represented in phase space. Its value does not vary with the continuous canonical coordinates,<ref group=note>Technically, there are some phases where the permutation of particles does not even yield a distinct specific phase: for example, two similar particles can share the exact same trajectory, internal state, etc.. However, in classical mechanics these phases only make up an infinitesimal fraction of the phase space (they have [[measure (mathematics)|measure]] zero) and so they do not contribute to any volume integral in phase space.</ref> so overcounting can be corrected simply by integrating over the full range of canonical coordinates, then dividing the result by the overcounting factor. However, {{math|''C''}} does vary strongly with discrete variables such as numbers of particles, and so it must be applied before summing over particle numbers.
| |
| | |
| As mentioned above, the classic example of this overcounting is for a fluid system containing various kinds of particles, where any two particles of the same kind are indistinguishable and exchangable. When the state is written in terms of the particles' individual positions and momenta, then the overcounting related to the exchange of identical particles is corrected by using<ref name="gibbs"/>
| |
| :<math>C = N_1! N_2! \ldots N_s!.</math>
| |
| This is known as "correct Boltzmann counting".
| |
| | |
| == Ensembles in statistics ==
| |
| :{{main|Principle of maximum entropy|Markov random field}}
| |
| The formulation of statistical ensembles used in physics has now been widely adopted in other fields, in part because it has been recognized that the [[canonical ensemble]] or [[Gibbs measure]] serves to maximize the entropy of a system, subject to a set of constraints: this is the [[principle of maximum entropy]]. This principle has now been widely applied to problems in [[linguistics]], [[robotics]], and the like.
| |
| | |
| In addition, statistical ensembles in physics are often built on a [[principle of locality]]: that all interactions are only between neighboring atoms or nearby molecules. Thus, for example, [[lattice model (physics)|lattice models]], such as the [[Ising model]], model [[ferromagnetic]] materials by means of nearest-neighbor interactions between spins. The statistical formulation of the principle of locality is now seen to be a form of the [[Markov property]] in the broad sense; nearest neighbors are now [[Markov blanket]]s. Thus, the general notion of a statistical ensemble with nearest-neighbor interactions leads to [[Markov random field]]s, which again find broad applicability; for example in [[Hopfield network]]s.
| |
| | |
| == Operational interpretation ==
| |
| | |
| In the discussion given so far, while rigorous, we have taken for granted that the notion of an ensemble is valid a priori, as is commonly done in physical context. What has not been shown is that the ensemble ''itself'' (not the consequent results) is a precisely defined object mathematically. For instance,
| |
| | |
| * It is not clear where this ''very large set of systems'' exists (for example, is it a [[gas in a box|''gas'' of particles inside a container]]?)
| |
| | |
| * It is not clear how to physically generate an ensemble.
| |
| | |
| In this section we attempt to partially answer this question.
| |
| | |
| Suppose we have a ''preparation procedure'' for a system in a physics
| |
| lab: For example, the procedure might involve a physical apparatus and
| |
| some protocols for manipulating the apparatus. As a result of this preparation procedure some system
| |
| is produced and maintained in isolation for some small period of time.
| |
| By repeating this laboratory preparation procedure we obtain a
| |
| sequence of systems ''X''<sub>1</sub>, ''X''<sub>2</sub>,
| |
| ....,''X''<sub>''k''</sub>, which in our mathematical idealization, we assume is an [[Infinity|infinite]] sequence of systems. The systems are similar in that they were all produced in the same way. This infinite sequence is an ensemble.
| |
| | |
| In a laboratory setting, each one of these prepped systems might be used as input
| |
| for ''one'' subsequent ''testing procedure''. Again, the testing procedure
| |
| involves a physical apparatus and some protocols; as a result of the
| |
| testing procedure we obtain a ''yes'' or ''no'' answer.
| |
| Given a testing procedure ''E'' applied to each prepared system, we obtain a sequence of values
| |
| Meas (''E'', ''X''<sub>1</sub>), Meas (''E'', ''X''<sub>2</sub>),
| |
| ...., Meas (''E'', ''X''<sub>''k''</sub>). Each one of these values is a 0 (or no) or a 1 (yes).
| |
| | |
| Assume the following time average exists:
| |
| :<math> \sigma(E) = \lim_{N \rightarrow \infty} \frac{1}{N} \sum_{k=1}^N \operatorname{Meas}(E, X_k) </math>
| |
| For quantum mechanical systems, an important assumption made in the
| |
| [[quantum logic]] approach to quantum mechanics is the identification of ''yes-no'' questions to the
| |
| lattice of closed subspaces of a Hilbert space. With some additional
| |
| technical assumptions one can then infer that states are given by
| |
| density operators ''S'' so that:
| |
| :<math> \sigma(E) = \operatorname{Tr}(E S). </math>
| |
| | |
| We see this reflects the definition of quantum states in general: A quantum state is a mapping from the observables to their expectation values.
| |
| | |
| ==See also==
| |
| * [[density matrix]]
| |
| * [[ensemble (fluid mechanics)]]
| |
| * [[phase space]]
| |
| * [[Liouville's theorem (Hamiltonian)]]
| |
| * [[ensemble average]]
| |
| * [[replication (statistics)]]
| |
| | |
| ==Notes==
| |
| {{reflist|group=note}}
| |
| | |
| ==References==
| |
| {{commons category|Statistical ensemble}}
| |
| | |
| {{reflist}}
| |
| | |
| {{Statistical mechanics topics}}
| |
| | |
| [[Category:Concepts in physics]]
| |
| [[Category:Statistical ensembles| ]]
| |
| [[Category:Philosophy of thermal and statistical physics]]
| |