Standard Error of Replicates, Each With Standard Error

Click For Summary
The discussion centers on the challenge of calculating the standard error for multiple simulations of a computer model, each producing its own means and standard errors. As the simulations progress, the standard errors decrease, indicating convergence to equilibrium. The main query is whether it's valid to average the standard errors from these replicates to obtain a more accurate estimate. Clarification is needed on the terms used, particularly "replicate," and how the data will be utilized in analysis. The conversation suggests that understanding the independence of the data is crucial for determining the legitimacy of averaging the standard errors.
chemaddict
Messages
15
Reaction score
0
So I have a situation that I keep confusing myself with:

I am running a computer simulation that, after a set number of iterations (a block) will output the running average (mean) of a property and the standard error based on the previous blocks. So, as the simulation runs, these standard errors tend to shrink because the simulation settles into its "equilibrium".

The problem is, I have 6 of these simulations (all the same, replicates), each with their own means and standard errors for each block during the simulation. I could easily find the mean and standard errors of the means, but what about the standard error associated with each replicate? Is there anyway to legitimately average those?

Thanks for any help.
 
Physics news on Phys.org
It's unlikely that people are going to understand exactly what you mean by "block" or "replicates" or what it means for 6 simulations to be "all the same, replicates". Whether it is legitimate to average a set of numbers will depend not only on what the numbers represent, but what you intend to do with the average. I suggest that you describe your work more clearly, including what you intend to do with any statistics that you compute. (Does "replicate" mean "replication"? I've most often heard (in the USA) people speak of "replications" of a simulation. I haven't heard the term "replicate" used.)
 
Sure, he is confusing. This is what I made of it:

He has 6 simulations.
One simulations has "X" number of iterations/blocks.
He has computed the mean and standard deviation of a property for each of the simulation.

What I cannot make out is "I could easily find the mean and standard errors of the means, but what about the standard error associated with each replicate? Is there anyway to legitimately average those?"
 
I'm sorry about the confusion. I don't mean to be patronizing in any way, but I think a simpler analogy might be much better.

Say, for instance, that I am interested in the speed a ball rolls down a large hill. Somehow, I can measure it every 1/2 second as it rolls. I then calculate a running average of all measurements as well as a running standard error.

The running standard error would obviously be undefined for the first measurement, but would get smaller and smaller as subsequent measurements were taken because N increases (assuming the ball maintained a fairly constant speed).

When the ball reaches the bottom, I then have running averages for the entire descent, as well as running standard errors associated with them.

Now, assume I wanted more confidence in that data, so I decide to measure the ball rolling down the hill a second time. Then a third, fourth, etc. up to six. Each time, I have running averages and standard errors.

I'm not just interested in the final speed based on the final measurements, but I am interested specifically in how that running mean and error change over time (i.e. how the mean "equilibrates" and how the error shrinks). Another assumption here is that all of these times matched up perfectly. For example, I have measurements for all six experiments at 0 sec, 1/2 sec, 1 sec, ... and I want to combine all six running means and running errors at each point in time along the descent. Combining the means seems easy (just take the mean of the means). However, is there a way to use the running standard errors from the six experiments for a given time along the descent? Can they combine in some way to give a better estimate of some true standard error (if that even makes sense)?

Again I'm sorry about any confusion, but I am a little confused myself.
Thanks.

P.S. A "replicate" (noun) is just another experiment under the same conditions (or as close as possible).
 
Last edited:
A ball rolling down a hill at a roughly constant speed may not get a good reception on a physics forum, but fortunately I lean more toward mathematics than physics.

Let's say the hill is bumpy, so the ball's constant average speed is due small accelerations and de-accelerations. We can even include friction, which will make things quite complicated since the "normal force" of the ball against the hill will vary.

Suppose we have 6 experiments of rolling the ball down the hill and record the (x,y) location of the ball as a function of time, so we have functions x = X_i(t) and y =Y_i(t) for i= 1,2,..6.

Lets focus on X(t). We can be interested in various parameters of the population of all possible replications of the experiment (not just the 6 that were done).

Examples of this are:

1) For a fixed t_{final} , what is the mean value of \frac{ X(t_{final}) - X(0} {t_{final} - 0}

2) For a fixed \triangle t and a randomly selected t , what is the mean value of \frac{X(t + \triangle t) - X(t)}{\triangle t }

3) For a given time t, what is the mean value of the deriviative \frac{dX}{dt} averaged over all possible replications of the experiment.

To decide the relevance of such parameters, we must decide what we are trying to accomplish. One can make deterministic representations of physical phenomena or stochastic models of physical phenomena.

A simple minded deterministic model for the phenomena would be to find a function F(t) such that F(t) is the parameter given by 3).

This establishes the relevance of parameter 3), but it leaves open the question of how to estimate it from the data. For example, to estimate F(t) at t = 9.4 should we average only the 6 velocities a time t= 9.4 or may we use other times or average velocities?

The answer to that depends on whether the other data can be regarded as independent (or approximately independent) random samples of the same random variable. You might be able to answer that question from your knowledge of the physics of a particular problem. If not, you can investigate the empircal evidence for dependence of different types data.

Perhaps you can elaborate further on this example to describe exactly what you final goal is.
 
The standard _A " operator" maps a Null Hypothesis Ho into a decision set { Do not reject:=1 and reject :=0}. In this sense ( HA)_A , makes no sense. Since H0, HA aren't exhaustive, can we find an alternative operator, _A' , so that ( H_A)_A' makes sense? Isn't Pearson Neyman related to this? Hope I'm making sense. Edit: I was motivated by a superficial similarity of the idea with double transposition of matrices M, with ## (M^{T})^{T}=M##, and just wanted to see if it made sense to talk...

Similar threads

  • · Replies 13 ·
Replies
13
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 25 ·
Replies
25
Views
3K
  • · Replies 25 ·
Replies
25
Views
12K
  • · Replies 21 ·
Replies
21
Views
3K
  • · Replies 19 ·
Replies
19
Views
2K
  • · Replies 9 ·
Replies
9
Views
4K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
2
Views
2K
  • · Replies 6 ·
Replies
6
Views
12K