Factorization Theorem for Sufficient Statistics & Indicator Function

Click For Summary

Discussion Overview

The discussion revolves around the application of the factorization theorem to demonstrate that the maximum of a random sample from a uniform distribution is a sufficient statistic for the parameter theta. Participants explore the properties of indicator functions and their dependence on the sample data and the parameter.

Discussion Character

  • Technical explanation
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • One participant questions the equality of the product of indicator functions for the sample values and the indicator function for the maximum value, suggesting that an additional factor is necessary to account for the smallest observation.
  • Another participant clarifies that the left side of the equation indicates that all sample values are within the interval [0, theta), which is equivalent to the maximum being within that interval.
  • A participant expresses uncertainty about whether the indicator function is a function of both the maximum value and theta, noting that it seems to depend only on the maximum value when theta is treated as a fixed parameter.
  • Further discussion raises the point that the definition of a function may need to consider constraints or conditions, suggesting that restrictions can influence the nature of the function.

Areas of Agreement / Disagreement

Participants express differing views on the relationship between the maximum and minimum observations in the context of the indicator functions, indicating that the discussion remains unresolved regarding the correct formulation of the factorization theorem in this case.

Contextual Notes

Participants highlight potential limitations in understanding the roles of the maximum and minimum observations in the context of the indicator functions, as well as the implications of treating parameters as fixed versus variable.

kingwinner
Messages
1,266
Reaction score
0
Problem:
Let Y1,Y2,...,Yn denote a random sample from the uniform distribution over the interval (0,theta). Show that Y(n)=max(Y1,Y2,...,Yn) is a sufficient statistic for theta by the factorization theorem.

Solution:
http://www.geocities.com/asdfasdf23135/stat10.JPG
1) While I understand that IA (x)IB (x)=IA intersect B (x), I don't understand the equality circled in red above.

In the solutions, they say that I0,theta (y1)...I0,theta (yn)=I0,theta (y(n)). Is this really correct?
Shouldn't the right hand side be I0,theta (y(n))I0,infinity (y(1)) ? I believe that the second factor is necessary because the largest observation is greater than zero does not guarantee that the smallest observation is greater than zero.
Which one is correct?


2) Also, is I0,theta (y(n)) a function of y(n), a function of theta, or a function of both y(n) and theta?
If it is a function of both y(n) and theta, then there is something that I don't understand. Following the definition of indicator function that IA (x) is a function of x alone (it is a function of only the stuff in the parenthesis), shouldn't I0,theta (y(n)) be a function of only y(n) alone?


Thank you for explaining! I've been confused with these ideas for at least a week.
 
Physics news on Phys.org
The left side of

<br /> \prod_{i=1}^n I_{[0,\theta)} (y_i) = I_{[0,\theta)} (y_{(n)})<br />

means that all the y_i values are in the interval [0,\theta).

This is true if, and only if, the maximum of the y's is in the same interval, and that is the meaning of the right-side.

The indicator I_{[0,\theta)} (y_{(n)}) is a function of y_{(n)} only, since \theta is fixed (it's a parameter).
 
I understand that
0<X_1,..., X_n<theta here these are the unordered data
is the same as (iff)
0<X_(1)<X_(2)<...<X_(n)<theta

But I don't think
0<X_(1)<X_(2)<...<X_(n)<theta
is EQUIVALENT to (iff)
0<X_(n)<theta
The => direction is true but <= is not. (the fact that the largest observation x(n) is greater than zero does not guarantee that the smallest observation x(1) is greater than zero.)

So that's why I think we should have
I0,theta(y1)...I0,theta(yn) = I0,theta(y(n))I0,infinity(y(1))
instead of I0,theta(y1)...I0,theta(yn)=I0,theta(y(n)).

Right?
 
Last edited:
statdad said:
The indicator I_{[0,\theta)} (y_{(n)}) is a function of y_{(n)} only, since \theta is fixed (it's a parameter).

But we can also write it as I y(n), inf (theta), in this case theta would be in the parenthesis, so in this case, would it be a function of theta alone? (in the gerenal case, f(x) means a function of x, f(y) means a function of y, the stuff in the parenthesis)


When we talk about functions, is it always only a function of the stuff in the parenthesis? It looks like that the restrictions/constraints are also important, so shouldn't it be a function also of the variables in the restrictions/constraints?

e.g.)
f(x)=x if x>y
f(x)=x^3 if x<y
Here not only the value of x controls f, the value of y also controls f, so is f a function of BOTH x and y here?

Thanks!
 

Similar threads

  • · Replies 6 ·
Replies
6
Views
2K
Replies
3
Views
2K
Replies
1
Views
1K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 11 ·
Replies
11
Views
4K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 4 ·
Replies
4
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K