MLE of μ for X1,...Xn with Known σ2i

In summary, it is possible to estimate all parameters of an n-observation (X1,...Xn) with the same mean, μ, but different variances (σ21,σ22,...,σ2n). The maximum likelihood estimator for μ is the standard weighted average, and the mle for σi2 is taken as exact. However, an estimator for the width of the distribution in μ can be found through error propagation in the equation for the mle of μ, using the associated σi as the error in xi.
  • #1
kimkibun
30
1
is it possible to estimate all parameters of an n-observation (X1,...Xn) with same mean, μ, but different variances (σ2122,...,σ2n)? if we assume that σ2i are known for all i in {1,...n}, what is the mle of of μ?
 
Physics news on Phys.org
  • #2
kimkibun said:
is it possible to estimate all parameters of an n-observation (X1,...Xn) with same mean, μ, but different variances (σ2122,...,σ2n)? if we assume that σ2i are known for all i in {1,...n}, what is the mle of of μ?

It is possible to do so, but you will have to weight the variances correctly for your estimator.

We know that E[Sum(Xi)/N] = E[mu] since all distributions have the same mean.

In terms of the variance we know that the variance of a normal estimator of a mean is sigma^2/n. Now because we can add variances and because all samples are independent, it means our final variance of the estimator is Sum (sigma_i^2/n_i) wher sigma_i is the standard deviation at index i and n_i is the number of samples from that particular distribution. If all sigma's are the same, this should simplify to sigma^2/N where N is the total number of samples (I'd double check for yourself).

So with that said our estimator will have the distribution (X - mean)/SQRT(Sum(sigma_i^2/n_i)) ~ N(0,1) under CLT assumption.

Now here is the thing though: if you are using classical statistical tests, you are going to assume the Central Limit Theorem which assumes you have enough samples for it to be normally distributed (or close enough to it) so that you can estimate the mean.

Because you have lots of sigma's, if you don't have enough samples with respect to the number of sigma's or if the sigma's are wildly different, then the assumption might break down.

There is even more to consider, but the above should give you an idea in terms of using just the population parameters of expectation.

If you are using things like t-tests, you should use the above kind of formulation and go from the derivation to include the above condition of how the variance of the estimator of the mean is the sum of all the variances. I'm guessing the S^2 will look similar but you'd have to make sure.
 
  • #3
i forgot to tell you that the n-observations was drawn out from a normal population..well anyway, is it possible that different variances might affect the maximum likelihood of the mean?
 
  • #4
kimkibun said:
i forgot to tell you that the n-observations was drawn out from a normal population..well anyway, is it possible that different variances might affect the maximum likelihood of the mean?

They do. If you go through the maximum likelihood logic with different variances, you'll find that the maximum likelihood estimator for the mean is the standard weighted average:
[tex]\frac{\sum_{i=1}^N w_i x_i}{\sum_{i=1}^N w_i}[/tex]
with
[tex]w_i = \frac{1}{\sigma_i^{\phantom{i}2}}.[/tex]
 
  • #5
Parlyne said:
They do. If you go through the maximum likelihood logic with different variances, you'll find that the maximum likelihood estimator for the mean is the standard weighted average:
[tex]\frac{\sum_{i=1}^N w_i x_i}{\sum_{i=1}^N w_i}[/tex]
with
[tex]w_i = \frac{1}{\sigma_i^{\phantom{i}2}}.[/tex]

can you please show me how you get that?

the mle for μ that i got is this,

(Ʃxi/Ʃσi2)(Ʃ1/σi2)

is it possible to find the mle of the parameter σi2?
 
  • #6
kimkibun said:
can you please show me how you get that?

the mle for μ that i got is this,

(Ʃxi/Ʃσi2)(Ʃ1/σi2)

is it possible to find the mle of the parameter σi2?

Presuming Gaussian errors, the probability of measuring a value within [itex]dx_i[/itex] of [itex]x_i[/itex] given mean [itex]\mu[/itex] and width [itex]\sigma_i[/itex] is
[tex]\frac{dx_i}{\sigma_i\sqrt{2\pi}}e^{-\frac{(x_i-\mu)^2}{2\sigma_i^{\phantom{i}2}}}.[/tex]

The probability of the full data set is the product of the probabilities for each value, which will be
[tex]\frac{\prod_{i=1}^N dx_i}{(2\pi)^\frac{N}{2}\prod_{i=1}^N \sigma_i} e^{-\frac{1}{2}\chi^2},[/tex]
with
[tex]\chi^2 = \sum_{i=1}^N\left(\frac{x_i-\mu}{\sigma_i}\right)^2.[/tex]

Since [itex]\chi^2[/itex] is the only part of the probability expression which depends on [itex]\mu[/itex], maximizing it over [itex]\mu[/itex] is equivalent to minimizing [itex]\chi^2[/itex]. So, we take the derivative of [itex]\chi^2[/itex] with respect to [itex]\mu[/itex] and set it to 0 as follows.
[tex]\begin{align} 0 &= \frac{d}{d\mu}\sum_{i=1}^N \frac{x_i^{\phantom{1}2}-2\mu x_i +\mu^2}{\sigma_i^{\phantom{i}2}}\\
&= \sum_{i=1}^N \frac{-2x_i+2\mu}{\sigma_i^{\phantom{i}2}}\\
&= 2 \mu \sum_{i=1}^N \frac{1}{\sigma_i^{\phantom{i}2}} - 2 \sum_{i=1}^N \frac{x_i}{\sigma_i^{\phantom{1}2}}\\
\mu &= \frac{\sum_{i=1}^N \frac{x_i}{\sigma_i^{\phantom{i}2}}}{\sum_{i=1}^N \frac{1}{\sigma_i^{\phantom{i}2}}}\end{align}[/tex]

As for [itex]\sigma_i[/itex], since there's a different value for each measurement, there's nothing to estimate. Each is taken as exact. What you can look for an estimator of is the width of the distribution in [itex]\mu[/itex], which can be found through error propagation in the equation for the mle of [itex]\mu[/itex], taking the associated [itex]\sigma_i[/itex] as the error in [itex]x_i[/itex].
 
  • #7
thank you so much sir! i really appreciate your efforts! God Bless you!
 

1. What is the MLE of μ for X1,...Xn with Known σ2i?

The maximum likelihood estimator (MLE) of μ for X1,...Xn with known σ2i is a statistic that estimates the true mean (μ) of a population based on a sample of n independent and identically distributed (i.i.d.) random variables (X1,...Xn) with known variance (σ2i). It is the value of μ that maximizes the likelihood function, which is a measure of how likely the observed data are given a specific value of the parameter μ.

2. How is the MLE of μ for X1,...Xn with Known σ2i calculated?

The MLE of μ for X1,...Xn with known σ2i is calculated by taking the derivative of the likelihood function with respect to μ, setting it equal to 0, and solving for μ. This results in the sample mean (X̄) being the MLE of μ, which is an unbiased estimator of μ. The formula for the MLE of μ is: μ^ = X̄ = (X1 + ... + Xn)/n.

3. What are the assumptions for using the MLE of μ for X1,...Xn with Known σ2i?

The assumptions for using the MLE of μ for X1,...Xn with known σ2i are that the sample of n i.i.d. random variables (X1,...Xn) is representative of the population, and that the population has a normal distribution with known variance (σ2i). If these assumptions are not met, the MLE of μ may not be the best estimator for the population mean.

4. How does the known variance (σ2i) affect the MLE of μ for X1,...Xn?

The known variance (σ2i) affects the MLE of μ for X1,...Xn by providing more information about the population distribution. This allows for a more precise estimate of the population mean, as the likelihood function is more sensitive to changes in μ when the variance is known. However, if the known variance is incorrect or does not accurately represent the population, the MLE of μ may be biased or inaccurate.

5. Can the MLE of μ for X1,...Xn with Known σ2i be used for any type of data?

No, the MLE of μ for X1,...Xn with known σ2i is specifically designed for continuous data that follows a normal distribution. It may not be appropriate for other types of data, such as categorical or discrete data, as the assumptions for using the MLE may not be met. In these cases, other estimation methods should be used to estimate the population mean.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
918
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
818
  • Set Theory, Logic, Probability, Statistics
Replies
23
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
16
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
762
  • Set Theory, Logic, Probability, Statistics
Replies
11
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
900
Back
Top