LBJking123
- 13
- 0
This is the question:
Suppose that X1...Xn form a random sample from the Bernoulli Distribution with unknown parameter P. Let Po and P1 be specified values such that 0<P1<Po<1, and suppose that is desired to test the following simple hypotheses: Ho: P=Po, H1: P=P1.
A. Show that a test procedure for which α(δ) + β(δ) is a minimum rejects Ho when Xbar < c.
B. Find the value of c.
I know that this problem is not that difficult I just can't figure out where to start. I know the Bernoulli distribution, but I can't figure out how to get α(δ) and β(δ). I have not seen any problems like this so I am kinda lost, any help would be much appreciated. Thanks!
Suppose that X1...Xn form a random sample from the Bernoulli Distribution with unknown parameter P. Let Po and P1 be specified values such that 0<P1<Po<1, and suppose that is desired to test the following simple hypotheses: Ho: P=Po, H1: P=P1.
A. Show that a test procedure for which α(δ) + β(δ) is a minimum rejects Ho when Xbar < c.
B. Find the value of c.
I know that this problem is not that difficult I just can't figure out where to start. I know the Bernoulli distribution, but I can't figure out how to get α(δ) and β(δ). I have not seen any problems like this so I am kinda lost, any help would be much appreciated. Thanks!