Thursday, June 28, 2012

Maximum Entropy Distributions

Entropy is an important topic in many fields; it has very well known uses in statistical mechanics, thermodynamics, and information theory. The classical formula for entropy is Σi(pi log pi), where p=p(x) is a probability density function describing the likelihood of a possible microstate of the system, i, being assumed. But what is this probability density function? How must the likelihood of states be configured so that we observe the appropriate macrostates?

In accordance with the second law of thermodynamics, we wish for the entropy to be maximized. If we take the entropy in the limit of large N, we can treat it with calculus as S[φ]=∫dx φ ln φ. Here, S is called a functional (which is, essentially, a function that takes another function as its argument). How can we maximize S? We will proceed using the methods of calculus of variations and Lagrange multipliers.

First we introduce three constraints. We require normalization, so that ∫dx φ = 1. This is a condition that any probability distribution must satisfy, so that the total probability over the domain of possible values is unity (since we’re asking for the probability of any possible event occurring). We require symmetry, so that the expected value of x is zero (it is equally likely to be in microstates to the left of the mean as it is to be in microstates to the right — note that this derivation is treating the one-dimensional case for simplicity). Then our constraint is ∫dx x·φ = 0. Finally, we will explicitly declare our variance to be σ², so that ∫dx x²·φ = σ².

Using Lagrange multipliers, we will instead maximize the augmented functional S[φ]=∫(φ ln φ + λ0φ + λ1xφ + λ2x²φ dx). Here, the integrand is just the sum of the integrands above, adjusted by Lagrange multipliers λk for which we’ll be solving.

Applying the Euler-Lagrange equations and solving for φ gives φ = 1/exp(1+λ0+xλ1+x²λ2). From here, our symmetry condition forces λ1=0, and evaluating the other integral conditions gives our other λ’s such that q = (1/2πσ²)½·exp(-x² / 2σ²), which is just the Normal (or Gaussian) distribution with mean 0 and variance σ². This remarkable distribution appears in many descriptions of nature, in no small part due to the Central Limit Theorem.

### Notes

1. rabi-goat-princess reblogged this from sayitwithscience
2. stochastastic reblogged this from sayitwithscience
3. marcospevensie reblogged this from sayitwithscience
4. jack-brain reblogged this from sayitwithscience
5. invictascientia reblogged this from sayitwithscience
6. megliotordichemais reblogged this from sayitwithscience
7. statempre reblogged this from sayitwithscience
8. phycafe reblogged this from sayitwithscience
9. darthbetty reblogged this from sayitwithscience
10. jonnynicknames reblogged this from sayitwithscience
11. another1one reblogged this from sayitwithscience
12. lillabet reblogged this from sayitwithscience
13. geometricespo reblogged this from sayitwithscience
14. ticktocktickertape reblogged this from sayitwithscience
15. starstuffblog reblogged this from sayitwithscience
16. gordanac reblogged this from sayitwithscience
17. quantumphysicshomebot reblogged this from sayitwithscience
18. vergiliusmaro reblogged this from sayitwithscience
19. youve-gotta-free-your-mind reblogged this from sayitwithscience and added:
Oh wow, just mixed APES with Physics and Calc. I love science/math.
20. tangereendream reblogged this from sayitwithscience
21. numadhamani reblogged this from sayitwithscience
22. afossilizedblog reblogged this from sayitwithscience and added:
Say It With Science is back! (I hope… *stares at SIWS members*) Here, I present a post written by my friend, Matthew.