A Correlation between parameters in a likelihood fit

Aleolomorfo
Messages
70
Reaction score
4
TL;DR Summary
Which is the right way to estimate the correlation between parameters estimated with a likelihood function?
Hello community!

I am facing a conceptual problem with the correlation matrix between maximum likelihood estimators.

I estimate two parameters (their names are SigmaBin0 and qqzz_norm_0) from a multidimensional likelihood function, actually the number of parameters are larger than the two I am focusing my attention now. I need to evaluate the correlation between that two parameters.
I know the best way to evaluate the correlation between two parameters from a likelihood fit is starting from the Hessian matrix. Performing that method I get a correlation of -0.14.

Then I tried a different approach: studying the SigmaBin0 vs qqzz_norm_0 values when one of the parameters is the Parameters Of Interest (POI) and the other is profiled, and viceversa. I mean, I scan the likelihood function along SigmaBin0 while "watching" the qqzz_norm_0 profiled values, and then I run another scan along qqzz_norm_0 while "watching" the SigmaBin0 profiled values. My expectation is to find the same trend in both cases, but what I find is the right plot in attachment. The vertical line is the former case, instead the horizontal line is the latter one. If I apply the definition of correlation (ratio of the covariance wrt the standard deviations) I get a correlation of -0.29.
On the other hand, if I consider the two trends independently and apply again the definition of correlation I got a correlation of -1 (central and left plot in attachment).

In my mind all these approaches should be equivalent and giving the same value of the correlation, but it is not the case. So there might be some bug in my reasoning. Can someone help me to sort out my ideas, please?
 

Attachments

  • Screenshot 2021-02-26 at 11.16.34.png
    Screenshot 2021-02-26 at 11.16.34.png
    17 KB · Views: 193
Physics news on Phys.org
Have you thought about using a Bayesian approach? With a Bayesian approach you get a posterior sample of the joint distribution of all of your parameters. So you can just directly calculate the correlation.

That said, I am not sure what you mean by:
Aleolomorfo said:
I scan the likelihood function along SigmaBin0 while "watching" the qqzz_norm_0 profiled values
So I cannot really assess what problems you run into. But this procedure does not seem clear to me so I am not surprised that it gives different results than the Hessian approach.
 
Aleolomorfo said:
I mean, I scan the likelihood function along SigmaBin0 while "watching" the qqzz_norm_0 profiled values, and then I run another scan along qqzz_norm_0 while "watching" the SigmaBin0 profiled values.

It isn't clear to me what you mean by "scan" and "profile". These are not standard terms in statistics although they may be familiar to people in your particular field of study.

I am facing a conceptual problem with the correlation matrix between maximum likelihood estimators.
For a "correlation between two estimators" to make sense, we have to define the estimators as random variables. We could imagine that you have multiple independent data sets ##D_1,D_2,D_3,...## and from each data set ##D_i## we get one pair of maxiumum likihood estimators ##(S_i, q_i)## (estimating different two parametrs). Then we can regard these pairs of values as random samples from two random variables. Is that what you are doing?
 
Stephen Tashi said:
For a "correlation between two estimators" to make sense, we have to define the estimators as random variables.
I agree. That is another reason to use the Bayesian approach here. This information is very natural and easy to obtain in that approach.
 
Hi all, I've been a roulette player for more than 10 years (although I took time off here and there) and it's only now that I'm trying to understand the physics of the game. Basically my strategy in roulette is to divide the wheel roughly into two halves (let's call them A and B). My theory is that in roulette there will invariably be variance. In other words, if A comes up 5 times in a row, B will be due to come up soon. However I have been proven wrong many times, and I have seen some...
Thread 'Detail of Diagonalization Lemma'
The following is more or less taken from page 6 of C. Smorynski's "Self-Reference and Modal Logic". (Springer, 1985) (I couldn't get raised brackets to indicate codification (Gödel numbering), so I use a box. The overline is assigning a name. The detail I would like clarification on is in the second step in the last line, where we have an m-overlined, and we substitute the expression for m. Are we saying that the name of a coded term is the same as the coded term? Thanks in advance.
Back
Top