Fisher matrix for multivariate normal distribution

AI Thread Summary
The Fisher information matrix (FIM) for a multivariate normal distribution can be expressed as \mathcal{I}_{m,n} = \frac{\partial \mu^\mathrm{T}}{\partial \theta_m} \Sigma^{-1} \frac{\partial \mu}{\partial \theta_n}, with derivation challenges noted when \Sigma depends on \theta. The initial derivation utilizes matrix derivatives, leading to a simplified form of the FIM, but difficulties arise in extending this to cases where \Sigma is parameter-dependent. References such as Porat & Friedlander's work and Klein & Neudecker's direct derivation provide insights into the FIM's computation. Additionally, there are formatting issues with TeX code that hinder clarity in the discussion. Overall, the quest for a comprehensive derivation continues, highlighting the complexity of the topic.
hdb
Messages
3
Reaction score
0
The fisher information matrix for multivariate normal distribution is said at many places to be simplified as:
\mathcal{I}_{m,n} = \frac{\partial \mu^\mathrm{T}}{\partial \theta_m} \Sigma^{-1} \frac{\partial \mu}{\partial \theta_n}.\
even on
http://en.wikipedia.org/wiki/Fisher_information#Multivariate_normal_distribution"
I am trying to come up with the derivation, but no luck so far. Does anyone have any ideas / hints / references, how to do this?

Thank you
 
Last edited by a moderator:
Physics news on Phys.org
Using matrix derivatives one has D_x(x^T A x) = x^T(A+A^T) from which it follows that D_{\theta} \log p(z ; \mu(\theta) , \Sigma) = (z-\mu(\theta))^T \Sigma^{-1} D_{\theta} \mu(\theta) For simplicity let's write D_{\theta} \mu(\theta) = H The FIM is then found as J = E[ ( D_{\theta} \log p(z ; \mu(\theta) , \Sigma))^T D_{\theta} \log p(z ; \mu(\theta) , \Sigma)] = E[ H^T R^{-1} (z - \mu(\theta))^T (z - \mu(\theta)) R^{-1} H] = H^T R^{-1} R R^{-1} H = H^T R^{-1} H [\tex] which is equivalent to the given formula. Notice that this formula only is valid as long as \Sigma [\tex] does not depend on \theta [\tex]. I'm still struggling to find a derivation of the more general case where also \Sigma [\tex] depends on \theta [\tex].<br /> <br /> For some reason my tex code is not correctly parsed. I cannot understand why.
 
Actually the general proof can apparently be found in Porat & Friedlander: Computation of the Exact Information Matrix of Gaussian Time Series with Stationary Random Components, IEEE Transactions on Acoustics, Speech and Signal Processing, Vol ASSP-34, No. 1, Feb. 1986.
 
edmundfo said:
R^{-1} H] = H^T R^{-1} R R^{-1} H = H^T R^{-1} H [\tex]

For some reason my tex code is not correctly parsed. I cannot understand why.

For one thing, you're using the back slash [\tex] instead of the forward slash [/tex] at the end of your code.
 
edmundfo said:
Actually the general proof can apparently be found in Porat & Friedlander: Computation of the Exact Information Matrix of Gaussian Time Series with Stationary Random Components, IEEE Transactions on Acoustics, Speech and Signal Processing, Vol ASSP-34, No. 1, Feb. 1986.
Thank you for the answers, in between I have found an another reference, which is a direct derivation of the same result, for me this one seems to be easier to interpret:

Klein, A., and H. Neudecker. “A direct derivation of the exact Fisher information matrix of Gaussian vector state space models.” Linear Algebra and its Applications 321, no. 1-3
 
I was reading documentation about the soundness and completeness of logic formal systems. Consider the following $$\vdash_S \phi$$ where ##S## is the proof-system making part the formal system and ##\phi## is a wff (well formed formula) of the formal language. Note the blank on left of the turnstile symbol ##\vdash_S##, as far as I can tell it actually represents the empty set. So what does it mean ? I guess it actually means ##\phi## is a theorem of the formal system, i.e. there is a...
Back
Top