Characteristic Function of Joint Gaussian Distribution

Jolb
Messages
417
Reaction score
29
This is inspired by Kardar's Statistical Physics of Particles, page 45, and uses similar notation.

Homework Statement


Find the characteristic function, \widetilde{p}(\overrightarrow{k}) for the joint gaussian distribution:

p(\overrightarrow{x})=\frac{1}{\sqrt{(2\pi)^{N}det<br /> [C]}}exp[-\frac{1}{2}\sum_{m, n}^{N}C_{mn}^{-1}(x_{m}-\lambda_{m})(x_{n}-\lambda_{n})]

Where C is a real symmetric matrix and C-1 is its inverse.
(Note that the -1 is an exponent, not subtraction of the identity matrix. Anytime I write X-1 I'm talking about the inverse of the matrix X).

Homework Equations


\widetilde{p}(\overrightarrow{k})=\int p(\overrightarrow{x})exp[-i\sum_{j=1}^{N}k_jx_j]d^{N}\overrightarrow{x}

That is, the characteristic function is the Fourier transform of the probability distribution.

The Attempt at a Solution


The first part of this problem was to find the normalization factor for the joint gaussian, which is the term in the square root in the expression for p(x). The way I did that was by noting that since C is real and symmetric, there must be an orthogonal matrix D such that (D-1)CD is diagonal with C's eigenvalues along the diagonal. Then, changing to the variables y_i = x_i - lambda_i, i.e. Dy=x-lambda, the sum over m,n reduces to just the sum over one index, and the integral breaks into the product of N gaussian integrals with variances equal to the eigenvalues of C.

That part was pretty confusing to me but fortunately Kardar tells you the recipe. He says, "The corresponding joint characteristic function is obtained by similar manipulations, and is given by

http://img194.imageshack.us/img194/9551/newfirst.png
."

Unfortunately I don't see how you can get rid of the extra x's when you try to perform the Fourier transform...

\widetilde{p}(\overrightarrow{k})=\int p(\overrightarrow{x})\exp[-i\sum_{j=1}^{N}k_jx_j]d^{N}\overrightarrow{x}
=\int_{\infty}^{\infty }\int_{\infty}^{\infty }...\int_{\infty}^{\infty }exp\left (\sum_{m, n}^{N}[C_{mn}^{-1}(x_{m}-\lambda_{m})(x_{n}-\lambda_{n})]-i\sum_{j}^{N}k_{j}x_{j} \right )dx_1dx_2...dx_N

Pretty ugly. Now if I try to change coordinates to the y's,
http://img715.imageshack.us/img715/4754/secondpv.png

Where the alphas are the eigenvalues of C and their product (of all N of them) is det[C].

This just doesn't seem to help. I'm really not sure what to do. Should I look for some other coordinates and find a new D to diagonalize? Or am I missing something? Is there some sort of orthogonality trick that let's me throw away a bunch of terms in the sum?

Any help would be greatly appreciated.

P.S. For some reason I couldn't get two of those equations to work in the PF markup. ?
 
Last edited by a moderator:
Physics news on Phys.org
I think writing things out in sums for this problem obscures what's really going on. I think keeping things as matrices and vectors makes them a lot clearer.

p({\bf x}) = (2\pi)^{N/2} |{\bf C}|^{-1} \exp\left(-\frac{1}{2}({\bf x}-{\bf \mu})^T {\bf C}^{-1}({\bf x}-{\bf \mu})\right)
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \int \exp\left(-\frac{1}{2}({\bf x}-{\bf \mu})^T {\bf C}^{-1}({\bf x}-{\bf \mu})+ i {\bf k}^T {\bf x}\right) d {\bf x}

Now, you correctly notices that
{\bf C} = {\bf D}{\bf \sigma}{\bf D}^{-1}
Where {\bf \sigma} is a diagonal matrix and {\bf D} is orthogonal. However did you know that {\bf D} being orthogonal means that
{\bf D}^{-1} = {\bf D}^{T}

Okay so define
{\bf y} = {\bf D}({\bf x}-{\bf \mu})
Hence
{\bf x} = {\bf D}^T{\bf y}+{\bf \mu}
And so after a change of variables in the integral we get
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \int \exp\left(-\frac{1}{2}{\bf y}^T {\bf \sigma}^{-1}{\bf y}+ i {\bf k}^T ({\bf D}^T{\bf y}+{\bf \mu}))\right) d {\bf y}
Pull out the mean
p({\bf k}) = (2\pi)^{N/2} |{\bf C}|^{-1} \exp( i {\bf k}^T{\bf \mu})\int \exp\left(-\frac{1}{2}{\bf y}^T {\bf \sigma}^{-1}{\bf y}+ i {\bf k}^T{\bf D}^T{\bf y})\right) d {\bf y}
Okay now hint: the {\bf D} in the integral is a rotation. Think about it rotating in k-space as opposed to real space. So define
{\bf \kappa} = {\bf D} {\bf k}
 
Thanks a lot Kajetan! The ideas of factoring out the means and rotating k were what was eluding me. I got tricked into following Kardar's notation of writing out the sums--I figured he wrote them like that as a hint!
 
Thread 'Need help understanding this figure on energy levels'
This figure is from "Introduction to Quantum Mechanics" by Griffiths (3rd edition). It is available to download. It is from page 142. I am hoping the usual people on this site will give me a hand understanding what is going on in the figure. After the equation (4.50) it says "It is customary to introduce the principal quantum number, ##n##, which simply orders the allowed energies, starting with 1 for the ground state. (see the figure)" I still don't understand the figure :( Here is...
Thread 'Understanding how to "tack on" the time wiggle factor'
The last problem I posted on QM made it into advanced homework help, that is why I am putting it here. I am sorry for any hassle imposed on the moderators by myself. Part (a) is quite easy. We get $$\sigma_1 = 2\lambda, \mathbf{v}_1 = \begin{pmatrix} 0 \\ 0 \\ 1 \end{pmatrix} \sigma_2 = \lambda, \mathbf{v}_2 = \begin{pmatrix} 1/\sqrt{2} \\ 1/\sqrt{2} \\ 0 \end{pmatrix} \sigma_3 = -\lambda, \mathbf{v}_3 = \begin{pmatrix} 1/\sqrt{2} \\ -1/\sqrt{2} \\ 0 \end{pmatrix} $$ There are two ways...
Back
Top