Complex conjugation and vector space duality

In summary: \rangle^*= c_1^* \left(c_1|\psi_1\right) + c_2^*...\right)^*=c_1^*\left(c_1|\psi_1\right)^*+c_2^*...\right)
  • #1
batboio
19
0
Hi all. First of all I want to say that I am new here and I want to apologise if this is the wrong forum to post my question. :)
I've just finished a sort of a crash course on functional analysis that left much more questions in me than it answered but one thing bothers me a lot.
When we talk about dual Hilbert spaces it seems that the vector in the dual space has coordinates that are simply complex conjugates of the coordinates of the corresponding vector.
Or using the Dirac notation:
|x>*=<x|, where x is from H - a Hilbert space
|x>=(x1, x2,...); <x|=(x1*, x2*,...)
First I want to ask if I got this right. And second - is there a way to generalise this for dual spaces as a whole (not just Hilbert spaces where H=H*). In other words is there any connection between complex conjugation and dual vectors?
 
Physics news on Phys.org
  • #2
There's a trick here. The Hilbert space must be complex and the scalar product must be a sequilinear form. Then, using Riesz' theorem and assuming separability of the HS, one shows that the action of a functional on a vector can be expressed as

[tex] F_{\phi} \psi = \langle \phi, \psi\rangle = \int_{\mathbb{R}^n} \phi^* (x) \psi (x) \, dx [/tex]
 
  • #3
batboio said:
|x>*=<x|, where x is from H - a Hilbert space
This is a really ugly notation. I wouldn't use it. I assume that it was inspired by the fact that some authors use x,y,z for members of H and x*,y*,z* for members of H*, but the * on x* is just there to remind the reader that we're dealing with a member of H*. x* is unrelated to x.
 
  • #4
Fredrik said:
x* is unrelated to x.

Isn't there a relation between the coordinates of a vector and a covector. In particular the coordinates of the covector in the basis of L*:
[tex]x_{i} = g_{ij} x^{j}[/tex],
where g is the metric tensor in L, [tex]x \in L[/tex]

Anyway this is the thing that bothers me exactly:
http://en.wikipedia.org/wiki/Bra-ket_notation#Hermitian_conjugation
It seems that after Hermitian conjugation the ket becomes a bra (which is essentially the vector becoming covector if I get it right).
 
  • #5
Fredrik said:
but the * on x* is just there to remind the reader that we're dealing with a member of H*. x* is unrelated to x.
I don't think so. The bra-ket formalism works (or at least the connection with inner products exists) because of Riesz Representation theorem: for a complex Hilbert space H, the map

[tex]H\to H^*[/tex]

[tex]h\mapsto \left<h,-\right>[/tex]

is an isometric anti-linear isomorphism. You can call this map (somewhat bad notation, agreed) *, so that [itex]h^*=\left<h,-\right>[/itex].

In bra-ket notation:

If [tex]h=\left|\phi\right>\in H[/tex], then [tex]h^*=\left|\phi\right>^*=\left<h,-\right>=\left<{\phi}\right|\in H^*[/tex].
 
  • #6
In the notation I described, x*,y*,z* are just symbols for arbitrary members of H*. I might as well have written α,β,γ.

Edit: A note for math guys like Landau: Inner products are usually taken to be linear in the second variable in physics books. I'm using that convention below.

I don't recall seeing the claim that the dagger operation takes a bra to the corresponding ket and vice versa before. I'm trying to see a "point" of that definition... The inner product satisfies <x,y>=<y,x>*. A bra acting on a ket is by definition the inner product of the kets, i.e. if I temporarily use (,) instead of <,> for inner products, we have

[tex]\langle\alpha|\beta\rangle =\langle\alpha||\beta\rangle=(|\alpha\rangle,|\beta\rangle)[/tex]

[tex]=(|\beta\rangle,|\alpha\rangle)^*=(\langle\beta||\alpha\rangle)^*[/tex]

For operators, we have the identity (AB)*=B*A*. If we insist that that should hold for bras and kets as well, the right-hand side above becomes

[tex]=|\alpha\rangle^*\langle\beta|^*[/tex]

which only makes sense if * changes bras to the corresponding kets and vice versa. I guess this is the "point" of that definition, but I don't see why we need (AB)*=B*A* to hold for bras and kets.

Let's see if this is consistent with the standard definition of the adjoint. A bra is a linear operator from H into ℂ, and ℂ is a Hilbert space. This means that * acting on a bra is already defined. The inner product on ℂ is defined by <z,w>=z*w. Let B be a bounded bra, i.e. let B:H→ℂ be linear and bounded. I'll use h for a member of H, and z for a member of ℂ. B*:ℂ→H is defined by

<z,Bh>=<B*z,h>

for all z and all h. So B* isn't a member of H or H**, as a ket should be, it's a function that takes complex numbers to kets. Let's continue anyway. The above means that the adjoint of a linear combination of bras can be evaluated like this

<(aA+bB)*z,h> = <z,(aA+bB)h> = a<z,Ah> + b<z,Bh> = a<A*z,h> + b<B*z,h>

= <a*A*z,h> + <b*B*z,h> = <(a*A*+b*B*),h>

This implies that (aA+bB)*=a*A*+b*B*. If we now use that other definition, we get the identity that Wikipedia wrote as

[tex]\left(c_1|\psi_1\rangle + c_2|\psi_2\rangle\right)^\dagger = c_1^* \langle\psi_1| + c_2^* \langle\psi_2|[/tex]
 
  • #7
Fredrik said:
I don't recall seeing the claim that the dagger operation takes a bra to the corresponding ket and vice versa before.
You haven't? What do/did you think the dagger operation 'is'? In my understanding, essentially daggering is just dualizing, the covariant functor which sends a vector space to its dual and maps to the corresponding dual map, but with two caveats: by Riesz, H and H* are canonically isomorphic, and we are working over the complex numbers so some complex scalars have to be replaced by their conjugate.

In R^n, this is just 'taking the transpose', so that the 'dagger' of a column vector is the corresponding row vector.
which only makes sense if * changes bras to the corresponding kets and vice versa. I guess this is the "point" of that definition, but I don't see why we need (AB)*=B*A* to hold for bras and kets.
Now I am even more interested: what is your (understanding of the) definition of * or dagger? I have always taken it to be nothing more than the Riesz isomorphism

[tex]H\to H^*[/tex]

[tex]h\mapsto \left<h,-\right>[/tex]
 
  • #8
Landau said:
Now I am even more interested: what is your (understanding of the) definition of * or dagger? I have always taken it to be nothing more than the Riesz isomorphism

[tex]H\to H^*[/tex]

[tex]h\mapsto \left<h,-\right>[/tex]
The first one, not the second. If you're used to seeing * for the second one too, I'm not going to argue with that. I just don't remember seeing it myself. If you have a notation like [itex]\langle\alpha|[/itex] at your disposal, you don't need the notation [itex]|\alpha\rangle^*[/itex] for the same thing.
 
  • #9
Fredrik said:
The first one, not the second.
What do you mean? The first what?
If you're used to seeing * for the second one too, I'm not going to argue with that. I just don't remember seeing it myself.
No no, my quantum mechanics classes are more than one year ago, I am not too sure what our notation back then was. I was just wondering what [itex]|\alpha\rangle^*[/itex] means to you. Now I understand that this just wasn't defined at all.

In any case, I think the whole formalism is consistent if we take this map to be the definition of *. If we want to extend the meaning of * to also apply to bra's, then it should assign to it the corresponding ket.
 
  • #10
The best way to understand this is to start with a complex vector space V sans the inner product. This vector space has a dual space V* of linear functionals mapping V to C. When V is finite dimensional V* is isomorphic as a vector space (has the same dimension) and even if not finite we can (e.g. with the inner product) define a corresponding isometric dual sub-space.

OK so given any basis [itex][e_1,e_2, e_3...][/itex] of V there is a dual basis [itex][f_1,f_2,f_3...][/itex] of V* so that
[tex]f_j(e_k)=\delta_{jk}[/tex]
that is each dual basis covector maps its corresponding basis vector to 1 and the others to 0. (Note that changing just one of the basis vectors can redefine all of the dual basis).

Now we can either define an inner product on V or a conjugation mapping from V to V*. The two are equivalent in that a conjugation defines an inner product and an inner product defines a conjugation.

Now defining a hermitian inner product [tex]\langle \quad,\quad\rangle[/tex] on V (thus making V a Hilbert space) the corresponding conjugate mapping, [tex]\dagger[/tex] maps an ortho-normal basis to its dual basis but it must be an ortho-normal basis.

So given a Hilbert space vector [tex]\psi\in V[/tex] the dual vector [tex]\psi^\dagger[/tex] (linear functional on V) is defined as the mapping:
[tex]\psi^\dagger(\phi) = \langle \psi , \phi \rangle[/tex]
or if you like this equation defines the inner product in terms of the dual mapping.

Now back to the Bra-Ket notation...
Given a vector [tex]\psi[/tex] and (ortho-normal) basis, [tex] e_1,e_2,\cdots[/tex] the ket [tex]|\psi\rangle[/tex] is the column vector of the basis expansion:
[tex]|\psi\rangle = (\psi_1,\psi_2,\cdots)^T\quad s.t. \quad \psi = \psi_1 e_1 + \psi_2 e_2+\cdots[/tex]

It then must be the case that the dagger dual of psi expanded in the dual basis will have complex conjugate expansion which we express as a row vector:
[tex]\langle \psi | = (\psi^*_1,\psi*_2,\cdots)[/tex]

We then, thus copy the dagger operation from the hilbert space in question into the row and column spaces of coefficients where it becomes the conjugate transpose. (That is to say we reuse the symbol.) We likewise have the inner product of two Hilbert space vectors in terms of the product of bra (row vector) with ket (column vector):
[tex] \langle \psi , \phi\rangle \to \langle \psi | \cdot |\phi \rangle \equiv \langle \psi|\phi\rangle[/tex]

And of course everything works nicely as well when we expand operators in the product basis defining their matrix representation. The adjoint operators likewise correspond to the conjugate transpose matrices (provided we work with proper bases.)

Note that once we have our inner product we can dispense with any direct reference to the dual space and dual basis and work always in the hilbert space itself. When we need a dual vector we simply use an appropriate vector and the inner product or what's better work wholly with the the column vectors of coefficients for a given basis and their conjugate transposes.

But IMNSHO to properly understand the notation it is crucial to first understand the relationship between the dagger, inner product and the dual space, and how we resolve these in terms of a basis.

Issues can surface that require careful parsing of the notation, specifically in relativistic quantum mechanics where e.g. spinors will reside not in a hilbert space but rather a space with an indefinite metric/inner product/adjoint (a pseudo-hilbert or philbert space).
 
Last edited:
  • #11
Landau said:
What do you mean? The first what?
Oops, sorry, I misread your post. I meant the first of the two math expressions I quoted, but I see now that they both represent the same thing. For some reason I interpreted the first expression as [itex]A\mapsto A^*[/itex] where A is a linear operator and A* its adjoint, as defined in e.g. Conway. That's what the * symbol means to me, and the [itex]\dagger[/itex] symbol means the same thing.

Landau said:
In any case, I think the whole formalism is consistent if we take this map to be the definition of *. If we want to extend the meaning of * to also apply to bra's, then it should assign to it the corresponding ket.
I agree.
 
  • #12
Thanks a lot for the answers I think I am starting to get this :)
 
  • #13
The Riesz representation theorem

I need to make a note of this somewhere, and it might as well be here. The first theorem is just a useful observation, and the second is the Riesz representation theorem, which guarantees the existence of a unique bra for each (normalizable) ket.

Theorem 1: Suppose that [itex]\mathcal H[/itex] is a complex Hilbert space, and that [itex]x_0\in\mathcal H[/itex]. Define [itex]\phi:\mathcal H\rightarrow\mathbb C[/itex] by [itex]\phi(x)=\langle x_0,x\rangle[/itex] for all [itex]x\in\mathcal H[/itex]. Then

(a) [itex]\phi[/itex] is linear.
(b) [itex]\phi[/itex] is bounded.
(c) [itex]\|\phi\|=\|x_0\|[/itex].
(d) [itex](\ker \phi)^\perp=\mathbb Cx_0[/itex].

Proof:

(a) The inner product is linear in the second variable.
(b) By the Cauchy-Schwarz inequality, [itex]|\phi(x)|\leq\|x_0\|\|x\|[/itex].
(c) Part (b) implies [itex]\|\phi\|\leq \|x_0\|[/itex] (because the supremum is the least upper bound). Since a supremum is an upper bound, we also have

[tex]\|\phi\|=\sup_{x\in\mathcal H}\frac{|\phi(x)|}{\|x\|}\geq \frac{|\phi(x_0)|}{\|x_0\|}=\|x_0\| [/tex]

(d) [tex]x\in\ker\phi\Leftrightarrow 0=\phi(x)=\langle x_0,x\rangle\Leftrightarrow x_0\perp x\Leftrightarrow x\in(\mathbb Cx_0)^\perp[/tex]
Since [itex]\ker\phi[/itex] is a closed linear subspace, this implies [itex](\ker\phi)^\perp=\mathbb C x_0[/itex].

Note that (a) and (b) imply that [itex]\phi\in\mathcal H^*[/itex].



Theorem 2: Suppose that [itex]\mathcal H[/itex] is a complex Hilbert space and that [itex]\phi:\mathcal H\rightarrow\mathbb C[/itex] is linear and bounded (i.e. that [itex]\phi\in\mathcal H^*[/itex]). Then there's a unique [itex]x_0\in\mathbb H[/itex] such that [itex]\phi(x)=\langle x_0,x\rangle[/itex] for all [itex]x\in\mathcal H[/itex].

Proof: There can be at most one [itex]x_0\in\mathcal H[/itex] with the required properties, because

[tex]\forall x\in H\quad \langle x_0,x\rangle=\langle x_0',x\rangle\ \Rightarrow\ \forall x\in H\quad \langle x_0-x_0',x\rangle=0[/tex]

[tex]\Rightarrow\ 0=\langle x_0-x_0',x_0-x_0'\rangle=\|x_0-x_0'\|^2\ \Rightarrow x_0'=x_0[/tex]

Theorem 1 tells us that if there's an [itex]x_0[/itex] with the required properties, then [itex]x_0\in(\ker\phi)^\perp[/itex] and [itex]\dim(\ker\phi)^\perp=1[/itex]. Because of this, our strategy to prove existence will be to pick an arbitrary member of [itex](\ker\phi)^\perp[/itex], and then show that we can choose [itex]x_0[/itex] to be a multiple of it. (Note that this doesn't mean that we have to use theorem 1 in this proof. It only provides the motivation for the first half of the trick we will use).

We need to consider the possibility that [itex](\ker\phi)^\perp=\{0\}[/itex]. Since [itex]\ker\phi[/itex] is a closed linear subspace, the projection theorem says that any [itex]x\in\mathbb H[/itex] can be uniquely expressed as [itex]x=y+z[/itex] where [itex]y\in\ker\phi[/itex] and [itex]z\in(\ker\phi)^\perp[/itex], so if [itex](\ker\phi)^\perp=\{0\}[/itex] we must have z=0, and this implies x=y and therefore [itex]\ker\phi=\mathcal H[/itex]. This implies [itex]\phi=0[/itex]. If this is the case, we can simply choose [itex]x_0=0[/itex].

If [itex](\ker\phi)^\perp\neq\{0\}[/itex], it contains a non-zero vector y. Now note that

[tex]x-\frac{\phi(x)}{\phi(y)}y\in\ker\phi[/tex]

Define [itex]z=y/\phi(y)[/itex] (just to simplify the following expression a little). We have

[tex]0=\langle z,x-\phi(x)z\rangle=\langle z,x\rangle-\phi(x)\|z\|^2[/tex]

[tex]\phi(x)=\left\langle\frac{z}{\|z\|^2},x\right\rangle[/tex]

and this means that we can choose [itex]x_0=z/\|z\|^2[/itex].



Theorem 3: The map [itex]B:\mathcal H\rightarrow H^*[/itex], defined by [itex]B(x)=\langle x,\cdot\rangle[/itex] for all [itex]x\in\mathcal H[/itex], is an antilinear bijection. (The expression [itex]\langle x,\cdot\rangle[/itex] denotes the map [itex]y\mapsto \langle x,y\rangle[/itex] from [itex]\mathcal H[/itex] into [itex]\mathbb C[/itex], and theorem 1 tells us that this map is a member of [itex]\mathcal H^*[/itex]).

Proof: For any [itex]a,b\in\mathbb C[/itex] and any [itex]x,y\in\mathcal H[/itex], we have

[tex]B(ax+by)z=\langle ax+by,z\rangle=a^*\langle x,z\rangle+b^*\langle y,z\rangle=(a^*B(x)+b^*B(y))z[/tex]

for all [itex]z\in\mathcal H[/itex], so B is antilinear. B is also injective, because

[tex]B(x)=B(y)\ \Rightarrow\ \forall z\in\mathcal H\quad \langle x,z\rangle=\langle y,z\rangle\ \Rightarrow\ \forall z\in\mathcal H\quad \langle x-y,z\rangle=0[/tex]

[tex]\Rightarrow\ 0=\langle x-y,x-y\rangle=\|x-y\|^2\ \Rightarrow\ x=y[/tex]

We don't need to prove that B is surjective here, because that's precisely what theorem 2 says, and we have already proved that.
 
Last edited:
  • #14
I forgot to prove that B (the antilinear bijection from [itex]\mathcal H[/itex] into [itex]\mathcal H^*[/itex]) is an isometry. This means that it's continuous as well. I should have made this a part of theorem 3.

[tex]\|B(x)\|=\sup_{y\in\mathcal H}\frac{|B(x)y|}{\|y\|}=\sup_{y\in\mathcal H}\frac{|\langle x,y\rangle|}{\|y\|}
\begin{cases}
{\displaystyle\leq\sup_{y\in\mathcal H}\frac{\|x\|\|y\|}{\|y\|}=\|x\|}\\
\\
{\displaystyle\geq \frac{|\langle x,x\rangle|}{\|x\|}=\|x\|}
\end{cases}
[/tex]
 
  • #15
Fredrik said:
Oops, sorry, I misread your post. I meant the first of the two math expressions I quoted, but I see now that they both represent the same thing.
Yes, the two 'expressions' formed the description of one map:

"the Riesz isomorphism
[tex]H\to H^*[/tex]

[tex]h\mapsto \left<h,-\right>[/tex]"
 

1. What is complex conjugation?

Complex conjugation is an operation that involves taking the complex conjugate of a complex number. This means changing the sign of the imaginary part of the number. For example, the complex conjugate of the number 3+4i is 3-4i.

2. How is complex conjugation related to vector space duality?

In vector space duality, the dual space is defined as the set of all linear functionals on a vector space. Complex conjugation can be thought of as a special type of linear functional that maps a complex number to its complex conjugate. Therefore, complex conjugation is closely related to vector space duality as it is an example of a linear functional.

3. What is the significance of complex conjugation in mathematics?

Complex conjugation is important in mathematics because it allows us to define the concept of orthogonality in complex vector spaces. It also plays a vital role in the study of complex analysis and is essential in solving certain types of equations in physics and engineering.

4. How does complex conjugation affect the properties of a vector space?

Complex conjugation does not change the underlying structure of a vector space, but it does affect the properties of the elements within that space. For example, the norm and inner product of a complex vector will change when complex conjugation is applied.

5. Can complex conjugation be extended to higher dimensional spaces?

Yes, complex conjugation can be extended to higher dimensional spaces. Just like in the 2-dimensional case, complex conjugation involves taking the complex conjugate of each component in a higher dimensional vector. This allows us to define orthogonality and other important concepts in these spaces.

Similar threads

  • Linear and Abstract Algebra
Replies
6
Views
863
  • Linear and Abstract Algebra
Replies
4
Views
2K
  • Linear and Abstract Algebra
Replies
3
Views
290
Replies
2
Views
1K
Replies
5
Views
1K
Replies
3
Views
1K
  • Linear and Abstract Algebra
Replies
21
Views
1K
  • Linear and Abstract Algebra
Replies
4
Views
2K
  • Linear and Abstract Algebra
Replies
13
Views
3K
Replies
2
Views
1K
Back
Top