Information stored in the initial condition of an ideal gas

Click For Summary

Homework Help Overview

The discussion revolves around estimating the amount of information required to store the initial conditions of gas particles, specifically in the context of statistical mechanics and information theory. The original poster presents a problem involving 1 mm³ of gas at normal conditions, containing approximately 1015 particles, and seeks to understand how to quantify the information associated with their positions and velocities.

Discussion Character

  • Exploratory, Conceptual clarification, Mathematical reasoning, Assumption checking

Approaches and Questions Raised

  • Participants explore different methods for calculating the information entropy of gas particles, considering both positions and velocities. Some suggest a simpler approach by focusing on the information needed to store the state of a single particle multiplied by the total number of particles. Others discuss the implications of using probability distributions and the nature of the initial conditions of the gas.

Discussion Status

The discussion is ongoing, with participants sharing various interpretations and approaches. Some guidance has been offered regarding the complexity of the problem, and there is a recognition of differing perspectives on how to quantify the information associated with the gas particles. The conversation reflects a mix of theoretical considerations and practical implications.

Contextual Notes

Participants note that the initial conditions of the gas are not probabilistic, as all particles have definite values, which raises questions about the necessity of statistical methods in this context. There is also mention of the challenge in measuring the state of the gas and how this relates to information entropy.

CY_Leung
Messages
6
Reaction score
0

Homework Statement


1 mm3 of gas at normal pressure and temperature contains about 1015 particles. Considering the particles as point-like and classical, provide a rough, conservative estimate for how many hard drives would be necessary to store the initial conditions of all gas particles. (As in 2013, a normal hard drive can store about 5 TB data.)

Homework Equations


This is the first course of statistical mechanics and actually we are not informed about the information theory in the class, except for the mathematical expression of information entropy. The following is all I know:
1 bit (1 Shannon) is defined as the amount of information of an event that occurs with a probability of 1/2.
Where 1 byte = 8 bits, 1 KB = 1000 bytes, 1 MB = 1000 KB, 1 GB = 1000 MB, 1 TB = 1000 GB
For an event with probability p, the number of bits of the corresponding amount of information is equal to -log2p.
Then the expected value of number of bits of a sample space is given by Σpilog2pi, which is actually the information entropy.
As for continuous probability distribution, one may write the information entropy as:
-∫p(x)log2p(x)dx

The Attempt at a Solution


Each classical and point-like particle can be fully described by (x,v) (or (x,p), doesn't really matter here). My interpretation of the question is: calculate the information entropy of the gas with the logarithmic base 2.
For information entropy is additive, i.e. for independent events, the total information entropy can be just summed.
I(p1p2) = I(p1) + I(p2) (read that from Wikipedia)
Thus velocities and positions of the particles can be considered separately, and each individual particle can also be considered separately.
First consider the velocity of a particle. What comes to my mind directly is the Maxwell-Boltzmann probability distribution. For ease of expression, the probability distribution can be written as:
p(v)=Cv2e-βv2, C being the normalization constant and β = m/2kT.
Do the integration:
I(\overrightarrow{v})=-C\int_{0}^{\infty}\overrightarrow{v}^{2}e^{-\beta\overrightarrow{v}^{2}}{log}_{2}(\overrightarrow{v}^{2}e^{-\beta\overrightarrow{v}^{2}})d\overrightarrow{v}<br /> =-\frac{C}{{ln}2}\int_{0}^{\infty}\overrightarrow{v}^{2}e^{-\beta\overrightarrow{v}^{2}}(2{ln}|\overrightarrow{v}|-\beta\overrightarrow{v}^{2}) d\overrightarrow{v}
For the gas is under normal condition, the velocities of the particles are not too large and we may omit the term ln|\overrightarrow{v}|. By this approximation the result of the integration would be 3/(2 ln2) according to my computation.
But I have no idea how to do this for particle positions. I attempted in the following way:
I(\overrightarrow{x})=-\int_{V}^{ } p(\overrightarrow{x})log_{2} p(\overrightarrow{x}) d\overrightarrow{x}
V being the volume bounded by the gas, p(\overrightarrow{x}) is the uniform probability density distribution. But now comes the problem: How should I do the logarithm of the uniform probability density distribution function, which is not dimensionless and depends on the unit we use for volume?
Or is the above approach correct at all? Thanks.
 
Physics news on Phys.org
My feeling is that you are overthinking it. I would take a much simpler approach: how much information is needed to store the state of one particle times the number of particles.
 
DrClaude said:
My feeling is that you are overthinking it. I would take a much simpler approach: how much information is needed to store the state of one particle times the number of particles.
This is actually also my approach, but since the positions and velocities of particles subject to some specific probability distribution, and the number of bits is related to the probability of a certain event, I have used the probability distribution to estimate the amount of information.
 
I mean, I can also take an approach like using the fact that, e.g. 1 byte is required to store a floating number, but I think the information entropy specifies the lower bound of information amount, and the way of computer to store a floating number is only one specific way of coding that may not be the most efficient when compared with the theoretical minimum?
I am not familiar with these stuff so please point out my mistakes if you spot them.
 
CY_Leung said:
This is actually also my approach, but since the positions and velocities of particles subject to some specific probability distribution, and the number of bits is related to the probability of a certain event, I have used the probability distribution to estimate the amount of information.
But the initial condition of the gas is not probabilistic. All particles have definite values. It is because of our inability to even list this initial condition that statistics are needed.
 
DrClaude said:
But the initial condition of the gas is not probabilistic. All particles have definite values. It is because of our inability to even list this initial condition that statistics are needed.
Okay I see what you mean. I interpreted the question as the following: The initial state of the gas is unknown, so if I have some way to measure all 6N values, how much information would I obtain (i.e. the decrease in information entropy = the original information entropy of the system, for there are no uncertainties about the gas anymore), and this has something to do with probability.
But if I change the question a little bit, like what is the amount of the information I obtain when I measure all states of the gas at an arbitrary time, does it make my approach correct?
 

Similar threads

Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
Replies
4
Views
5K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 19 ·
Replies
19
Views
3K
Replies
1
Views
3K
  • · Replies 27 ·
Replies
27
Views
5K