Characteristic Function of Joint Gaussian Distribution

Click For Summary
SUMMARY

The discussion focuses on deriving the characteristic function, \(\widetilde{p}(\overrightarrow{k})\), for the joint Gaussian distribution defined by the probability density function \(p(\overrightarrow{x})\). The solution involves recognizing that the characteristic function is the Fourier transform of the probability distribution. Key steps include changing variables to simplify the integral and utilizing properties of orthogonal matrices to diagonalize the covariance matrix \(C\). The final expression for the characteristic function incorporates the mean vector and the eigenvalues of the covariance matrix.

PREREQUISITES
  • Understanding of joint Gaussian distributions and their properties
  • Familiarity with Fourier transforms in the context of probability distributions
  • Knowledge of matrix operations, specifically eigenvalues and orthogonal matrices
  • Experience with change of variables in multiple integrals
NEXT STEPS
  • Study the derivation of the Fourier transform of probability distributions
  • Learn about the properties of orthogonal matrices and their applications in diagonalization
  • Explore the implications of eigenvalues in multivariate statistics
  • Investigate advanced topics in statistical physics related to Gaussian distributions
USEFUL FOR

Mathematicians, physicists, statisticians, and data scientists interested in multivariate analysis and the applications of Gaussian distributions in statistical modeling.

Jolb
Messages
417
Reaction score
29
This is inspired by Kardar's Statistical Physics of Particles, page 45, and uses similar notation.

Homework Statement


Find the characteristic function, \widetilde{p}(\overrightarrow{k}) for the joint gaussian distribution:

p(\overrightarrow{x})=\frac{1}{\sqrt{(2\pi)^{N}det<br /> [C]}}exp[-\frac{1}{2}\sum_{m, n}^{N}C_{mn}^{-1}(x_{m}-\lambda_{m})(x_{n}-\lambda_{n})]

Where C is a real symmetric matrix and C-1 is its inverse.
(Note that the -1 is an exponent, not subtraction of the identity matrix. Anytime I write X-1 I'm talking about the inverse of the matrix X).

Homework Equations


\widetilde{p}(\overrightarrow{k})=\int p(\overrightarrow{x})exp[-i\sum_{j=1}^{N}k_jx_j]d^{N}\overrightarrow{x}

That is, the characteristic function is the Fourier transform of the probability distribution.

The Attempt at a Solution


The first part of this problem was to find the normalization factor for the joint gaussian, which is the term in the square root in the expression for p(x). The way I did that was by noting that since C is real and symmetric, there must be an orthogonal matrix D such that (D-1)CD is diagonal with C's eigenvalues along the diagonal. Then, changing to the variables y_i = x_i - lambda_i, i.e. Dy=x-lambda, the sum over m,n reduces to just the sum over one index, and the integral breaks into the product of N gaussian integrals with variances equal to the eigenvalues of C.

That part was pretty confusing to me but fortunately Kardar tells you the recipe. He says, "The corresponding joint characteristic function is obtained by similar manipulations, and is given by

http://img194.imageshack.us/img194/9551/newfirst.png
."

Unfortunately I don't see how you can get rid of the extra x's when you try to perform the Fourier transform...

\widetilde{p}(\overrightarrow{k})=\int p(\overrightarrow{x})\exp[-i\sum_{j=1}^{N}k_jx_j]d^{N}\overrightarrow{x}
=\int_{\infty}^{\infty }\int_{\infty}^{\infty }...\int_{\infty}^{\infty }exp\left (\sum_{m, n}^{N}[C_{mn}^{-1}(x_{m}-\lambda_{m})(x_{n}-\lambda_{n})]-i\sum_{j}^{N}k_{j}x_{j} \right )dx_1dx_2...dx_N

Pretty ugly. Now if I try to change coordinates to the y's,
http://img715.imageshack.us/img715/4754/secondpv.png

Where the alphas are the eigenvalues of C and their product (of all N of them) is det[C].

This just doesn't seem to help. I'm really not sure what to do. Should I look for some other coordinates and find a new D to diagonalize? Or am I missing something? Is there some sort of orthogonality trick that let's me throw away a bunch of terms in the sum?

Any help would be greatly appreciated.

P.S. For some reason I couldn't get two of those equations to work in the PF markup. ?
 
Last edited by a moderator:
Physics news on Phys.org
I think writing things out in sums for this problem obscures what's really going on. I think keeping things as matrices and vectors makes them a lot clearer.

p({\bf x}) = (2\pi)^{N/2} |{\bf C}|^{-1} \exp\left(-\frac{1}{2}({\bf x}-{\bf \mu})^T {\bf C}^{-1}({\bf x}-{\bf \mu})\right)
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \int \exp\left(-\frac{1}{2}({\bf x}-{\bf \mu})^T {\bf C}^{-1}({\bf x}-{\bf \mu})+ i {\bf k}^T {\bf x}\right) d {\bf x}

Now, you correctly notices that
{\bf C} = {\bf D}{\bf \sigma}{\bf D}^{-1}
Where {\bf \sigma} is a diagonal matrix and {\bf D} is orthogonal. However did you know that {\bf D} being orthogonal means that
{\bf D}^{-1} = {\bf D}^{T}

Okay so define
{\bf y} = {\bf D}({\bf x}-{\bf \mu})
Hence
{\bf x} = {\bf D}^T{\bf y}+{\bf \mu}
And so after a change of variables in the integral we get
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \int \exp\left(-\frac{1}{2}{\bf y}^T {\bf \sigma}^{-1}{\bf y}+ i {\bf k}^T ({\bf D}^T{\bf y}+{\bf \mu}))\right) d {\bf y}
Pull out the mean
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \exp( i {\bf k}^T{\bf \mu})\int \exp\left(-\frac{1}{2}{\bf y}^T {\bf \sigma}^{-1}{\bf y}+ i {\bf k}^T{\bf D}^T{\bf y})\right) d {\bf y}
Okay now hint: the {\bf D} in the integral is a rotation. Think about it rotating in k-space as opposed to real space. So define
{\bf \kappa} = {\bf D} {\bf k}
 
Thanks a lot Kajetan! The ideas of factoring out the means and rotating k were what was eluding me. I got tricked into following Kardar's notation of writing out the sums--I figured he wrote them like that as a hint!
 

Similar threads

  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 36 ·
2
Replies
36
Views
6K
Replies
2
Views
2K
Replies
2
Views
2K
  • · Replies 15 ·
Replies
15
Views
2K
Replies
1
Views
2K
Replies
1
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K