So if I have a logistic regression: ##log (\hat {odds})=\hat{\beta_{0}}+\hat{\beta_{1}}x##. How would I find a confidence interval for x if I am given ##odds=5## This is going in reverse, where if I have the outcome, I try to do inference on the predictor.(adsbygoogle = window.adsbygoogle || []).push({});

We know that ##\hat{\vec{\beta}}##, the vector of ##\hat{\beta_{0}}## and ##\hat{\beta_{1}}##, is distributed asmptotically as ##N( \vec{\beta}, \hat{{Var{\hat{\beta}}}})## for large sample sizes where ##\hat{{Var{\hat{\beta}}}}## is the estimated variance covariance matrix.( similar to how ##\sigma^{2}## is estimated by ##s^{2}## for the basic t test).

So solving for x using ##log (5)=\beta_{0}+\beta_{1}x## I get: ##x( \beta_{0},\beta_{1}) =\frac{ log(5)-\beta_{0}} {\beta_{1} }## a multivariate function of both parameters.

So it seems that I would use the delta method but the one problem is the variance covariance matrix is an estimated one, so it doesn't have any actual parameters in it. Would delta method still work for this? Or would I need to try another method?

**Physics Forums | Science Articles, Homework Help, Discussion**

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

# A Does Delta Method work for asymptotic distributions?

Have something to add?

Draft saved
Draft deleted

**Physics Forums | Science Articles, Homework Help, Discussion**