Unitary Operator Inverse Transformation: Solving for a_m

  • Thread starter Hazzattack
  • Start date
  • Tags
    Operators
In summary, the person is trying to figure out if what they are doing is legit. They ask if a unitary transform can be made between two sets of operators, and are asking for pointers on how to isolate a given matrix element. They state that if this is entirely wrong, some pointers on how to isolate a given matrix element would be appreciated. They ask for help with understanding the summation convention for matrix multiplication. They state that if linear algebra is not their strong point, some online material or literature on the subject may be helpful. Finally, they ask for help with understanding the transformation T given that T is linear and has an inverse.
  • #1
Hazzattack
69
1
Hi everyone, I was hoping that someone might be able to tell me if what I'm doing is legit.

Firstly, I start by saying that a unitary transform can be made between two sets of operators (this is defined in this specific way);

b[itex]_{n}[/itex] = [itex]\Sigma_{m}[/itex] U[itex]_{mn}[/itex]a[itex]_{m}[/itex] (1)

Now this is the bit I'm not sure about, if i want to do the inverse to solve for a[itex]_{m}[/itex] is the following acceptable;

[itex]\Sigma_{n}[/itex]U[itex]_{nm}[/itex]^{dagger}b[itex]_{n}[/itex] = [itex]\Sigma_{n}[/itex][itex]\Sigma_{m}[/itex]U[itex]_{mn}[/itex]U[itex]_{nm}[/itex]^{dagger}a[itex]_{m}[/itex]

Where [itex]\Sigma_{n}[/itex][itex]\Sigma_{m}[/itex]U[itex]_{mn}[/itex]U[itex]_{nm}[/itex]^{dagger} = (UU^{dagger})[itex]_{nn}[/itex]= [itex]\delta_{nn}[/itex] = 1

Thanks in advance.

If this is entirely wrong, some pointers on how I isolate a[itex]_{m}[/itex] would be appreciated.

In essence what I'm asking is given (1), what is the inverse transformation of it?

Extra information: The components of U[itex]_{mn}[/itex] are real.
 
Last edited:
Physics news on Phys.org
  • #2
In this paper - http://arxiv.org/abs/1006.4507

It seems to suggest that the inverse transform would be;

[itex]\Sigma_{n}[/itex] U[itex]_{nm}[/itex] b[itex]_{n}[/itex] = a[itex]_{m}[/itex]

I'm not entirely sure how, specifically the sum index is what is confusing me.
 
  • #3
The summation calls for summing over all values of n for each index value of m.

Thus you end up with a collection of items a[m].
 
  • #4
Yeah, i understand what the summation means. But i don't see how it follows from starting with (1), mathematically i mean. Could you possibly elaborate (by example)?

Thanks.
 
  • #5
Hazzattack said:
[itex]\Sigma_{n} U_{nm}^\dagger b_{n} = \Sigma_{n} \Sigma_{m} U_{mn} U_{nm}^{\dagger}a_{m}[/itex]

This is where you start going wrong. On the lhs, m is a free index, but on the rhs it was already a summation index from the previous eqn.

Try using $$ \Sigma_n U_{kn}^\dagger b_n ~=~ \cdots$$instead.

BTW, you've got some of your indices roung the wrong way. E.g., It's less error-prone to write
$$ b_n ~=~ \Sigma_m U_{nm} a_m$$because then it matches how "matrix ##\times## column vector" works.

You might also want to learn the summation convention to reduce the clutter.

Oh, and you don't need quite so many itex pairs in your equations. I adjusted the quoted part above to illustrate.
 
  • #6
Why don't you think about it in terms of transformations on basis vectors? I'm going to appeal to your training in linear algebra...

Let's say that you know some vector [itex]A[/itex] is related to another vector [itex]B[/itex] by some transformation [itex]T[/itex].

[itex]\vec{B}=T\vec{A}[/itex]

Now Let's assume that [itex]T[/itex] is linear and has an inverse. Multiply on the left by its inverse

[itex]T^{-1} T \vec{A} = T^{-1}\vec{B}[/itex]

[itex]\mathbb{1} \vec{A} = T^{-1}\vec{B}[/itex], where [itex]\mathbb{1}[/itex] is the unit operator

[itex]\vec{A} = T^{-1}\vec{B}[/itex]

we have now inverted the equation in a basis free way.

Now consider that we could express the vectors and the transformation in a specific basis and evaluate the transformation action the vector in the same way we do matrix multiplication. This is what you wrote down in your post, basically consider [itex]a_n[/itex] the basis vectors of [itex]\vec{A}[/itex] and [itex]b_n[/itex] the basis vectors of [itex]\vec{B}[/itex], for example.

So the only problem is finding the inverse of the transformation. But you know your transformation is unitary, so the inverse is the Hermitian conjugate [itex]U^{-1} = U^{\dagger}[/itex] so you could immediately write down

[itex]a_{n} = \sum_{m} U_{nm} b_{m}[/itex]
[itex]b_{n} = \sum_{m} (U^{\dagger})_{nm} a_{m}[/itex]

I hope this explanation is clear enough.

In case you really wanted to do it with a specific basis you'd do as strangerep suggested...
[itex]a_{n} = \sum_{n}U_{nm}b_{m}[/itex]
[itex]\sum_{k}(U^{\dagger})_{kn}a_{n} = \sum_{k}\sum_{n}(U^{\dagger})_{kn}U_{nm}b_{m}[/itex]
[itex]\sum_{k}(U^{\dagger})_{kn}a_{n} = \sum_{k}(U^{\dagger}U)_{km}b_{m}[/itex] ← If this step seems fishy, I could attempt to explain in a later post
[itex]\sum_{k}(U^{\dagger})_{kn}a_{n} = \sum_{k}δ_{km}b_{m}[/itex] (the Kronecker delta gives exactly the entries of the matrix form of [itex]\mathbb{1} = diag(1, 1, 1, \cdots, 1)[/itex])
[itex]\sum_{k}(U^{\dagger})_{kn}a_{n} = b_{k}[/itex]

But you could just re label the indices in the last equation.
[itex]b_{n} = \sum_{m} (U^{\dagger})_{nm} a_{m}[/itex]

I didn't really want to outright tell you everything since it might be homework, but I think it's important to understand that it's "only" linear algebra.
 
Last edited:
  • #7
Firstly, I want to thank both of you for taking the time to respond to me in the depth that you did. Both posts were very helpful.

Secondly, I'm working on this paper as part of my dissertation hence why I'm desperately trying to get to grips with all the notation. Could either (or both) of you perhaps suggest some online material/literature that may help with me getting up to speed on the all the index notations as I believe currently this is my largest source of errors (Just to put the paper into context they are using orthogonal polynomials because of their obvious benefits and I'm trying to replicate the method using a matrix representation).

Finally, i can follow everything you did Wotanub except for this line;

[itex]\sum_{k}(U^{\dagger})_{kn}a_{n} = b_{k}[/itex]

Why does the sum over K get dropped here (on the RHS)? Is it because every entry along the diagonal will equal 1?

Also, the index of b changes to k but on your first line it has the index m - what does this actually mean? As i would have thought you want the index to remain as m in order for it to be an inverse transformation (or are you free to change this index without any consequences).

I think what would be helpful is if I could find something with strict definitions in this summation notation. As I'm sure it's the actual notation that is confusing me the most.

Added: Do you think you could explain the whole free index thing? I'm a little confused when I'm supposed to use this.

Thanks again!
 
Last edited:
  • #8
Also... When you take the Hermitian conjugate of a matrix do you just flip the indices (since that is what a transpose does), such that;

[itex] (U^{\dagger})_{nm}[/itex] = [itex] (U)_{mn} [/itex]

Since in this specific case all entries of the unitary operator are real.
 
  • #9
I thought I should elaborate as you guys took time to respond and it may be obvious to you where I'm explicitly going wrong.

so in the paper they implement a transform such that;

b[itex]_{n}= \int dkh(k) \pi_{n}(k)a(k)[/itex] (where [itex]\pi_{n}(k)[/itex] is a normalised nth monic orthogonal polynomial and both this and h(k) are real)

I have said this can be equivalently written;

[itex] b_{n} = \Sigma_{k} h_{k} \pi_{nk} a_{k} [/itex] (1)

In this case [itex]U^{\dagger}_{nk} = h_{k} \pi_{nk} [/itex]

Therefore, we can write (1) as such;

[itex] b_{n} = \Sigma_{k} U^{\dagger}_{nk} a_{k} [/itex] (hence the first question i initially asked)

Now, it states in the paper that the inverse transformation can be written;

a(k) = [itex] \Sigma_{n=0}^{n=\infty} h(k) \pi_{n}(k)b_{n} [/itex]

Thus, I would have expected that in 'my' notation this should read;

[itex] a_{k} = \Sigma_{n} h_{k} \pi_{kn} b_{n} = \Sigma_{n} U_{kn} b_{n} ?? [/itex]

This is why I was asking the questions about the unitary operator as I can't seem to understand how to get it to that from starting from expression (1). I assume it's because I'm struggling with the index notation.

Thanks as always for taking the time to read this.
 
Last edited:
  • #10
Let me illustrate it correctly, because in the wotanub's post there are some index mistakes...
[itex] a_{n}=\sum_{m} U_{nm} b_{m} [/itex] Watch? I have a free index n on the lhs, so I keep it in the rhs too... the summation takes place over m...
[itex] \sum_{n} (U^{\dagger})_{kn} a_{n}=\sum_{m} \sum_{n} (U^{\dagger})_{kn} U_{nm} b_{m}[/itex]
[itex] \sum_{n} (U^{\dagger})_{kn} a_{n}=\sum_{m} (U^{\dagger}U)_{km} b_{m}[/itex]
[itex] \sum_{n} (U^{\dagger})_{kn} a_{n}=\sum_{m} δ_{km} b_{m}[/itex]
[itex] \sum_{n} (U^{\dagger})_{kn} a_{n}= b_{k}[/itex]
That is now fine... on the lhs there is only one free k, and I have the k on the rhs...n indices are summed over...
One point that might be sneaky is the where I erased the sum of n on the rhs...but that happened because the matrices multiplication is: [itex]A_{ab}= \sum_{r} B_{ar}C_{rb}[/itex]

Now, why is the sum of the delta of kroenicker with b give that result?
Because you have the property:
[itex]δ_{ab}=1 [/itex] for [itex]a=b[/itex] or [itex]δ_{ab}=0[/itex] else
So you can see it right away that from the sum over m, only the m=k will survive... The beginner's way would be to expand the sum:
[itex]\sum_{m} δ_{km} b_{m}=δ_{k1} b_{1}+δ_{k2} b_{2}+...+δ_{kN} b_{N}[/itex]
where I said that the sum happens from m=1 to m=N.
so if k=1, you will get only:
[itex]\sum_{m} δ_{km} b_{m}=δ_{11} b_{1}+δ_{12} b_{2}+...+δ_{1N} b_{N}=δ_{11} b_{1}=b_{1}[/itex]
if k=2, you'll get only [itex]b_{2}[/itex] in exactly the same way...
Finally for any k, you will get only the k-th component... So:
[itex]\sum_{m} δ_{km} b_{m}= b_{k}[/itex]


The free index trick is no trick. You don't use it, it exists... Indices in general can represent vector components. If you have free indices you have something like vectors. If you don't have those indices, then you have something like numbers/scalars. As such you cannot have that a vector is equal to a number, so if you have a free index on one side, you must also have it in the other side as well...
 
Last edited:
  • Like
Likes 1 person
  • #11
Thank you so much for responding. That all makes a lot of sense and was SO helpful. I see so in that case the Kroenicker delta is 'picking' out a single entity.

One quick question, the transpose of a unitary operator with only real entries is the same as just flipping the indices right? (since for a vector you are just going from for example a row vector to a column vector).

Does this imply you can write;

[itex] U^{\dagger}_{nm} = U_{mn} [/itex]
 
Last edited:
  • #12
If the unitary matrix is real, yes you can make it...
in general the dagger symbol means that:
[itex](U_{ab})^{\dagger}= (U_{ab}^{*})^{T}[/itex]
Now if the unitary matrix elements are real, you get immediately the transpose...
[itex](U_{ab})^{\dagger}= (U_{ab})^{T}[/itex]
 
  • #13
Great. I think I've obtained what I was after. Could you verify that I haven't added something (because I want it to be that way...) I shouldn't have.

I start by saying;

[itex] b_n = \Sigma_{m} U_{nm} a_{m} [/itex]

Thus, b=Ua taking the transpose of each side (all entries of U are real)

[itex] \Sigma_{n} U^{\dagger}_{kn}b_{n} = \Sigma_{m} \Sigma_{n} U^{\dagger}_{kn} U_{nm} a_{m} = \Sigma_{m} (U^{\dagger} U)_{km} a{m} = \Sigma_{m} δ_{km}a_{m} = a_{k}[/itex]

Since [itex] U^{\dagger}_{kn} = U_{nk} [/itex], the final expression yields;

[itex] \Sigma_{n} U_{nk} b_{n} = a_{k} [/itex]

This is the only way I can seem to make both derivations equivalent, otherwise I fall into problems when i sub back in for [itex]U_{nk} [/itex] etc (see above for definition). as [itex] h_{k} [/itex]is a diagonal matrix which should have the same index regardless.

Thanks again guys... Sorry I'm bombarding the forums...
 
  • #14
Hazzattack said:
I think what would be helpful is if I could find something with strict definitions in this summation notation. As I'm sure it's the actual notation that is confusing me the most.
Did you carefully study the Wikipedia link I gave you?

(Sorry, I'm pressed for time right now. I might try to reply further later if anything's still unclear. I'm glad ChrisVer corrected the typos in wotanub's post.)
 
  • #15
Hazzattack said:
I think what would be helpful is if I could find something with strict definitions in this summation notation. As I'm sure it's the actual notation that is confusing me the most.
I think you should start by taking another look at the definition of matrix multiplication. I will denote the entry on row i, column j of an arbitrary matrix X by ##X_{ij}##. The definition of matrix multiplication says that if A is an m×n matrix, and B is an n×p matrix, then AB is the m×p matrix such that for all ##i\in\{1,\dots,m\}## and all ##j\in\{1,\dots,p\}##,
$$(AB)_{ij}=\sum_{k=1}^n A_{ik}B_{kj}.$$ The summation convention, or at least the simplest version of it, is just the convention to drop the summation sigma from the notation and write the right-hand side as ##A_{ik}B_{kj}## instead.

This convention is based on the observation that the summation is always over all possible values of the index that appears twice, and there's (almost) always a sum over an index that appears twice. There are exceptions, like when you want to refer to the entry on row i, column i of A. If you have said that you're using the summation convention, you can't just write it as ##A_{ii}##, because this is interpreted as ##\sum_i A_{ii}=\operatorname{Tr} A##.
 

1. What is a unitary operator inverse transformation?

A unitary operator inverse transformation is a mathematical operation that reverses the effects of a unitary operator, which is a linear transformation on a vector space that preserves the length and inner product of vectors. In simpler terms, it is a way to "undo" a transformation and return to the original state.

2. How do you solve for a_m in a unitary operator inverse transformation?

To solve for a_m, you need to use the unitary property of the transformation, which states that the inverse of a unitary operator is equal to its adjoint (conjugate transpose). This means that to solve for a_m, you need to take the adjoint of the transformation and then apply it to the vector in question. The resulting vector will have the coefficients a_m that you are looking for.

3. Can a unitary operator inverse transformation be used for any vector space?

Yes, a unitary operator inverse transformation can be used for any finite-dimensional vector space. This is because the unitary property of the transformation holds for any vector space, as long as the inner product is defined.

4. What is the significance of solving for a_m in a unitary operator inverse transformation?

Solving for a_m allows us to find the coefficients of a vector in a different basis, which can be useful in many applications. For example, in quantum mechanics, solving for the coefficients in the energy eigenbasis allows us to determine the probability amplitudes of a particle being in each energy state.

5. Are there any limitations to using unitary operator inverse transformation to solve for a_m?

One limitation is that it can only be used for finite-dimensional vector spaces. Additionally, the unitary operator must be well-defined and invertible. There may also be computational limitations, depending on the size and complexity of the vector space and transformation.

Similar threads

Replies
1
Views
754
Replies
10
Views
1K
Replies
17
Views
1K
Replies
5
Views
983
  • Quantum Physics
Replies
1
Views
1K
  • Beyond the Standard Models
Replies
3
Views
2K
  • Quantum Physics
Replies
1
Views
1K
  • Quantum Physics
Replies
19
Views
3K
Replies
2
Views
1K
Back
Top