Usuário(a):MGromov/Testes34



Página inicial: Testes

Edições completas: Mecânica estatística  • Modelo Hodgkin-Huxley  • Neurociência computacional  • Modelo probabilístico para redes neurais  • Teoria de campo médio  • Modelo FitzHugh–Nagumo  • Processo Lévy  • Cadeias de Markov  • Processo de Poisson  • Galves–Löcherbach model  • Stochastic chains with memory of variable length  • Lesão do plexo braquial  • Somatotopia  • Função densidade  • Modelos de grafos aleatórios exponenciais • Processo de Gram-Schmidt  • Equação de Chapman–Kolmogorov  • Predefinição:Processos estocásticos  • Algoritmo de autovalores  • Transição de fase  • Hipótese do cérebro crítico  • Critical brain hypothesis  • Passeio aleatório  • Plasticidade sináptica  • Potencial pós-sináptico excitatório  • Potencial pós-sináptico inibitório  • Modelo de Morris-Lecar  • Plexo braquial  • Processo gaussiano  • Campo aleatório de Markov  • Eletroencefalografia  • Modelo de Hindmarsh-Rose  • Sistemas de partícula em interação  • Medida de Gibbs  • Nervo escapular dorsal  • Nervo radial  • Nervo peitoral lateral  • Nervo musculocutâneo  • Medida de Dirac  • Nervo torácico longo  • Sigma-álgebra  • Nervo peitoral medial  • Nervo ulnar  • Potencial evocado  • Estimulação magnética transcraniana repetitiva  • Teorema de Donsker  • Desigualdade de Boole  • Codificação neural  • Aprendizado de máquina  • Independência condicional  • Inferência estatística  • Nervo subclávio  • Nervo supraescapular  • Nervo mediano  • Nervo axilar  • Movimento browniano geométrico  • Caminho autoevitante  • Tempo local  • Nervo subescapular superior  • Nervo toracodorsal  • Nervo subscapular inferior  • Caminho (teoria dos grafos)  • Campo aleatório  • Lei do logaritmo iterado

Edições em andamento: Nervo cutâneo medial do braço (N)  • Nervo cutâneo medial do antebraço (N)  • Cérebro estatístico (N)  • Statistician brain  • Distribuição de probabilidade condicional (N)  • Esperança condicional (N)  • Integral de Itō (N)  • Martingale  • Variação quadrática (N)  • Processo Ornstein–Uhlenbeck  • Ruído branco  • Teoria ergódica  • Avalanche neuronal (N)  • Teoria da percolação (N)  • Função totiente de Euler  • Ruído neuronal (N)  • Distribuição de Poisson  • Córtex cerebral  • Estímulo (fisiologia)




In mathematics, the Gibbs measure, named after Josiah Willard Gibbs, is a probability measure frequently seen in many problems of probability theory and statistical mechanics. It is a generalization of the canonical ensemble to infinite systems. The canonical ensemble gives the probability of the system X being in state x (equivalently, of the random variable X having value x) as

Here, E(x) is a function from the space of states to the real numbers; in physics applications, E(x) is interpreted as the energy of the configuration x. The parameter β is a free parameter; in physics, it is the inverse temperature. The normalizing constant Z(β) is the partition function. However, in infinite systems, the total energy is no longer a finite number and cannot be used in the traditional construction of the probability distribution of a canonical ensemble. Traditional approaches in statistical physics studied the limit of intensive properties as the size of a finite system approaches infinity (the thermodynamic limit). When the energy function can be written as a sum of terms that each involve only variables from a finite subsystem, the notion of a Gibbs measure provides an alternative approach. Gibbs measures were proposed by probability theorists such as Dobrushin, Lanford, and Ruelle and provided a framework to directly study infinite systems, instead of taking the limit of finite systems.

A measure is a Gibbs measure if the conditional probabilities it induces on each finite subsystem satisfy a consistency condition: if all degrees of freedom outside the finite subsystem are frozen, the canonical ensemble for the subsystem subject to these boundary conditions matches the probabilities in the Gibbs measure conditional on the frozen degrees of freedom.

The Hammersley–Clifford theorem implies that any probability measure that satisfies a Markov property is a Gibbs measure for an appropriate choice of (locally defined) energy function. Therefore, the Gibbs measure applies to widespread problems outside of physics, such as Hopfield networks, Markov networks, Markov logic networks, and bounded rational potential games in game theory and economics. A Gibbs measure in a system with local (finite-range) interactions maximizes the entropy density for a given expected energy density; or, equivalently, it minimizes the free energy density.

The Gibbs measure of an infinite system is not necessarily unique, in contrast to the canonical ensemble of a finite system, which is unique. The existence of more than one Gibbs measures is associated with statistical phenomena such as symmetry breaking and phase coexistence.

Markov property

editar

An example of the Markov property can be seen in the Gibbs measure of the Ising model. The probability for a given spin σk to be in state s could, in principle, depend on the states of all other spins in the system. Thus, we may write the probability as

 .

However, in an Ising model with only finite-range interactions (for example, nearest-neighbor interactions), we actually have

 ,

where Nk is a neighborhood of the site k. That is, the probability at site k depends only on the spins in a finite neighborhood. This last equation is in the form of a local Markov property. Measures with this property are sometimes called Markov random fields. More strongly, the converse is also true: any positive probability distribution (nonzero density everywhere) having the Markov property can be represented as a Gibbs measure for an appropriate energy function.[1] This is the Hammersley–Clifford theorem.

Formal definition on lattices

editar

What follows is a formal definition for the special case of a random field on a lattice. The idea of a Gibbs measure is, however, much more general than this.

The definition of a Gibbs random field on a lattice requires some terminology:

  • The lattice: A countable set  .
  • The single-spin space: A probability space  .
  • The configuration space:  , where   and  .
  • Given a configuration ω ∈ Ω and a subset  , the restriction of ω to Λ is  . If   and  , then the configuration   is the configuration whose restrictions to Λ1 and Λ2 are   and  , respectively. These will be used to define cylinder sets, below.
  • The set   of all finite subsets of  .
  • For each subset  ,   is the σ-algebra generated by the family of functions  , where  . This σ-algebra is just the σ-algebra of cylinder sets on the lattice.
  • The potential: A family   of functions ΦA : Ω → R such that
    1. For each   is  -measurable.
    2. For all   and ω ∈ Ω, the following series exists:
 

We interpret ΦA as the contribution to the total energy (the Hamiltonian) associated to the interaction among all the points of finite set A. Then   as the contribution to the total energy of all the finite sets A that meet  . Note that the total energy is typically infinite, but when we "localize" to each   it may be finite, we hope.

  • The Hamiltonian in   with boundary conditions  , for the potential Φ, is defined by
 
where  .
  • The partition function in   with boundary conditions   and inverse temperature β > 0 (for the potential Φ and λ) is defined by
 
where
 
is the product measure
A potential Φ is λ-admissible if   is finite for all   and β > 0.
A probability measure μ on   is a Gibbs measure for a λ-admissible potential Φ if it satisfies the Dobrushin–Lanford–Ruelle (DLR) equation
 
for all   and  .

An example

editar

To help understand the above definitions, here are the corresponding quantities in the important example of the Ising model with nearest-neighbor interactions (coupling constant J) and a magnetic field (h), on Zd:

  • The lattice is simply  .
  • The single-spin space is S = {−1, 1}.
  • The potential is given by
 

See also

editar

References

editar
  1. Ross Kindermann and J. Laurie Snell, Markov Random Fields and Their Applications (1980) American Mathematical Society, ISBN 0-8218-5001-6

Further reading

editar
  • Georgii, H.-O. (2011) [1988]. Gibbs Measures and Phase Transitions 2nd ed. Berlin: de Gruyter. ISBN 978-3-11-025029-9 

Predefinição:Stochastic processes

Category:Measures (measure theory) Category:Statistical mechanics