Can I omit using an indicator function when estimating an MLE?

  • Thread starter Thread starter cdux
  • Start date Start date
  • Tags Tags
    Function Mle
Click For Summary
Using an indicator function is common when estimating the maximum likelihood estimator (MLE) for the uniform distribution, as it accounts for the non-zero probability region. However, it is possible to derive the MLE without it by focusing on boundaries and using order statistics, specifically the maximum order statistic. The indicator function method may seem redundant since both approaches yield similar conclusions regarding the parameter estimation. The order statistic method is often viewed as more intuitive, leveraging the distribution's characteristics. Ultimately, multiple estimation methods exist, and the choice depends on the estimator's effectiveness and personal preference.
cdux
Messages
187
Reaction score
0
When looking for a maximum likelihood estimator for the Uniform distribution I noticed that a common method is to use an indicator function. My initial understanding is that the reason for that is for taking into account the region of ℝ that x produces - or not - a non-zero probability.

If I go on finding the maximum likelihood function without involving an indicator function in the math and then in the end only mention the boundaries of effect of the result, is it correct?

I'd be happy if I can avoid it (correctly) because I find the indicator function method very unintuitive for my style.
 
Physics news on Phys.org
Hey cdux.

The uniform distribution can't use derivatives because the PDF is flat and also because the range of the PDF depends on the actual parameter (unlike the Normal distribution and others where it doesn't).

So for this particular case, you need to use either the indicator function or the order statistic to derive the estimator for the parameter of the uniform distribution.

If you are using the order statistic, then make sure you check the bias (if any iexists). Also note that when I say order statistic I mean the maximum order statistic X(n) where n is the sample size and X(n) is the nth order statistic distribution.
 
  • Like
Likes 1 person
I've found both types of solutions, and I have a hard time understanding the difference between the "easy" one (the 'ordered statistic' one) and the one with the indicator function: In both cases they seem to conclude with the same thing: "Since b ['a' was zero in that case but it applies for both], since b >= x by our initial assumptions, then it comes to reason that -n/b^n which is a decreasing function, must be satisfied with b = max{X1,X2,...Xn}".

Because that's what I was already doing, evaluating l'(b), then looking at it being a decreasing function and then concluding that b should be minimized to maximize L(b) which is our goal.

Unless.. the solutions with indicator functions I've been reading had been redundant and they "talked too much" so to speak. Because if you are going to explain the order of the random variables why use an indicator function to begin with?
 
Last edited:
The indicator formulation is something that can be maximized in the context of the MLE (which is all about maximizing some quantity).

The order statistics on the other hand is an intuitive method that uses the nature of the distribution and knowing that the last value is a good estimate of the actual parameter.

So in short, MLE looks at maximizing something in general and order statistic is an intuitive estimator based on the specific nature of the uniform distribution.

Whatever estimator is good is what you use and there are always multiple estimators for a single parameter (MLE, moment estimator, non-parametric etc).
 
The standard _A " operator" maps a Null Hypothesis Ho into a decision set { Do not reject:=1 and reject :=0}. In this sense ( HA)_A , makes no sense. Since H0, HA aren't exhaustive, can we find an alternative operator, _A' , so that ( H_A)_A' makes sense? Isn't Pearson Neyman related to this? Hope I'm making sense. Edit: I was motivated by a superficial similarity of the idea with double transposition of matrices M, with ## (M^{T})^{T}=M##, and just wanted to see if it made sense to talk...

Similar threads

  • · Replies 19 ·
Replies
19
Views
2K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 5 ·
Replies
5
Views
4K
Replies
3
Views
3K
  • · Replies 23 ·
Replies
23
Views
4K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 8 ·
Replies
8
Views
2K
  • · Replies 7 ·
Replies
7
Views
3K