Coulomb potential as an operator

In summary: Anyway, you should have already found (around post #7) that $$[p, A(x)] = \frac{d}{dx}A(x)$$ for any polynomial ##A(x)##. So, given ##F(x) = A^{-1}(x)##, what does that imply?
  • #1
ShayanJ
Insights Author
Gold Member
2,810
604
I want to calculate the commutator ##{\Large [p_i,\frac{x_j}{r}]}## but I have no idea how I should work with the operator ##{\Large\frac{x_j}{r} }##.
Is it ## x_j \frac 1 r ## or ## \frac 1 r x_j ##? Or these two are equal?
How can I calculate ##{\Large [p_i,\frac 1 r]}##?
Thanks
 
Physics news on Phys.org
  • #2
My first thought is that the momentum operator act as a derivative on a function of the operator r
 
  • #3
Since the ##x_j## commute and ##r^2=\sum x_j^2##, ##x## and ##r## commute. You can differentiate the relation defining ##r## to find the partial derivatives, and then use ##[p_k,f(x)]=-i\hbar \partial f(x)/\partial x_k##. (Don't use ##i## as index if it appears as ##\sqrt{-1}##, too.)
 
Last edited:
  • #4
I've been able to prove ## [p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k} ## only for the case of ## f(x) ## being a polynomial in ## x_k ##. But here ## \frac 1 r ## doesn't satisfy this condition. Is there a general proof?
 
Last edited:
  • #5
Shyan said:
##[p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k}##
Actually the differential operator should be of partial one.
Shyan said:
I've been able to prove ## [p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k} ## only for the case of ## f(x) ## being a polynomial in ## x_k ##. But here ## \frac 1 r ## doesn't satisfy this condition. Is there a general proof?

I have no idea what you're talking about!
The commutator suggested by Neumaier is applicable to any smooth function ##f(x)##, not only polynomials.
 
  • #6
blue_leaf77 said:
The commutator suggested by Neumaier is applicable to any smooth function
## f(x) ## is a function of positon operator, not a smooth function of a real variable!
 
  • #7
Who said that ##f(x)## is a function of real variable?
 
  • #8
blue_leaf77 said:
Who said that ##f(x)## is a function of real variable?
I got the impression that you meant it was trivial since you didn't provide any proof, so I thought that's what you mean. Do you know where can I find a proof?
 
  • #9
Consider
$$
[p,x^n] = -i\hbar nx^{n-1}
$$
and the Taylor expansion of ##f(x)##.
 
  • #10
blue_leaf77 said:
Consider
$$
[p,x^n] = -i\hbar nx^{n-1}
$$
and the Taylor expansion of ##f(x)##.

A while ago I asked about this and micromass didn't seem to agree with you.(See here!)
Actually that's why I didn't think about using the formula in the first place.
 
  • #11
It seems that Taylor expansion of a function operator cannot be taken lightly, I have actually also addressed exactly the same problem on this regard almost a year ago There I asked if a function ##f(x)=(1+x)^{-1}## can be expanded in power terms, as the same function with ##x## real numbers can only have Taylor expansion for a bounded x ##-1<x<1##.
 
  • #12
blue_leaf77 said:
It seems that Taylor expansion of a function operator cannot be taken lightly, I have actually also addressed exactly the same problem on this regard almost a year ago There I asked if a function ##f(x)=(1+x)^{-1}## can be expanded in power terms, as the same function with ##x## real numbers can only have Taylor expansion for a bounded x ##-1<x<1##.
Is there any proof that doesn't use Taylor expansion of operator functions?
 
  • #13
Shyan said:
Is there any proof that doesn't use Taylor expansion of operator functions?
Well, it starts from Taylor expansion of ordinary functions. But yes, one can extend to very general (complex analytic) functions.

To get started, restrict to the case where an (ordinary, polynomial) operator ##A## has an inverse ##A^{-1}##. I.e., ##A A^{-1} = 1##. Then take the commutator with ##p## on both sides, and use the Leibniz rule on the LHS. What do you find? :oldwink:[And thank you to @A. Neumaier for teaching me this generalized material ages ago.]
 
  • Like
Likes blue_leaf77
  • #14
strangerep said:
To get started, restrict to the case where an (ordinary, polynomial) operator A has an inverse A^{-1}. I.e., A A^{-1} = 1. Then take the commutator with p on both sides, and use the Leibniz rule on the LHS. What do you find? :oldwink:
For any operator B, I find ## [B,A]A^{-1}=A[A^{-1},B] ##! So?
 
  • #15
Shyan said:
For any operator B, I find ## [B,A]A^{-1}=A[A^{-1},B] ##! So?
Take it further...

When ##A = A(x)## (ordinary polynomial), and ##B = p##, and you've already found that ##[p, A(x)] = A'(x)## (modulo factors of ##i## and ##\hbar##, depending on your conventions),... then what is $$[p, A^{-1}(x)] ~=~ ?$$
 
  • #16
strangerep said:
Take it further...

When ##A = A(x)## (ordinary polynomial), and ##B = p##, and you've already found that ##[p, A(x)] = A'(x)## (modulo factors of ##i## and ##\hbar##, depending on your conventions),... then what is $$[p, A^{-1}(x)] ~=~ ?$$
I don't see how ## [p,A]A^{-1}=A[A^{-1},p] ## implies ##[p,A(x)]\propto A'(x)##!
 
  • #17
Shyan said:
I don't see how ## [p,A]A^{-1}=A[A^{-1},p] ## implies ##[p,A(x)]\propto A'(x)##!
You already had that back in post #4. Now use that to find ##[p,A^{-1}] = ?##
 
  • #18
## [p,A(x)^{-1}]=i\hbar A(x)^{-1} A'(x) A^{-1}(x)##
 
  • #19
You keep stopping too soon.

Let ##F(x) := A^{-1}(x)##. From the above, what is ##[p, F(x)] = \; ?##
 
  • #20
Can you be sure that A^-1 commutes with A' ?
 
  • #21
andresB said:
Can you be sure that A^-1 commutes with A' ?
They're both functions of x so I think yes, which gives ## [p,A^{-1}(x)]=i\hbar [A(x)^{-1}]^2 A'(x) ##.

strangerep said:
You keep stopping too soon.

Let ##F(x) := A^{-1}(x)##. From the above, what is ##[p, F(x)] = \; ?##
So I have ## [p,F(x)]=i\hbar [F(x)]^2 A'(x) ##. But what is ##A'(x)## in terms of ##F(x)##?
 
  • #22
andresB said:
Can you be sure that A^-1 commutes with A' ?
Sorry -- I should have specified that ##A^{-1}## is 2-sided inverse operator. But in the cases we're considering, where ##A## is a function of a single operator ##x##, it's a reasonably safe assumption, at least for purposes of physics (with appropriate caveats on the domain of definition).
 
  • #23
Shyan said:
So I have ## [p,F(x)]=i\hbar [F(x)]^2 A'(x) ##. But what is ##A'(x)## in terms of ##F(x)##?
Oh well, I guess we can all be a bit blind sometimes. Certainly, it happens to me.

You have essentially proven that $$[p, A^{-1}(x)] ~=~ \frac{d A^{-1}(x)}{dx} ~.$$ This generalizes the earlier derivative formula to a larger class of functions than mere polynomials.

From this, you can go further, to functions of the form ##f(x) = g(x)/h(x)##, and generalize the derivative formula even further.
 
  • Like
Likes ShayanJ
  • #24
Shyan said:
## [p,A(x)^{-1}]=i\hbar A(x)^{-1} A'(x) A^{-1}(x)##
strangerep said:
You have essentially proven that ##[p, A^{-1}(x)] ~=~ \frac{d A^{-1}(x)}{dx}##. This generalizes the earlier derivative formula to a larger class of functions than mere polynomials.

From this, you can go further, to functions of the form ##f(x) = g(x)/h(x)##, and generalize the derivative formula even further.
In more explicit terms: Once you know that ##[p,f(x)]=-i\hbar f'(x)## for some function ##f## of an operator vector ##x## with commuting components you get it for ##g(x):=f(x)^{-1}## in place of ##f(x)##, too. By linear combination if you start with ##f(x)=x^n## with a multiexponent ##n## (where it follows from the definition by induction) you get it first for all polynomials, then for their inverses, then for all rational functions. Then for limits of rational functions, for integrals, and Cauchy's integral theorem gives it for all functions analytic on the joint spectrum of ##x## (with exception of any singularities). This is what you need to handle the Coulomb potential.
 
Last edited:
  • Like
Likes ShayanJ and blue_leaf77
  • #25
A. Neumaier said:
In more explicit terms: [...]
Yes -- that's where I was heading. But I didn't want to "deprive" Shyan of the enjoyment I experienced (some time ago) of discovering these things by working them out. :biggrin:
 
  • #26
strangerep said:
Yes -- that's where I was heading. But I didn't want to "deprive" Shyan of the enjoyment I experienced (some time ago) of discovering these things by working them out. :biggrin:
I think he'll still enjoy it with my shortcut guide rather than a long interactive guide. There are enough steps to fill in (and small pitfalls) to make it mathematically rigorous.

Actually, one can do ##f=1/r## in an direct way by finding the implicit algebraic equation it satisfies in terms of ##x##, and proceeding similar to the case of the inverse of a polynomial. This is enjoyable, too, and one doesn't need limits. This works for any algebraic function, but not for exponentials, etc..
 
  • #27
I have to wait for 11 months to find a satisfying answer to this problem of reciprocal operator.
 
  • #28
Shyan said:
Is there any proof that doesn't use Taylor expansion of operator functions?

Just a suggestion. Look up the Stone-Weierstrass theorems. The simplest one works for real continuous functions over a real interval. But others work for complex functions, quaternions, etc. For the real case, "every continuous function defined on a closed interval[a, b] can be uniformly approximated as closely as desired by a polynomial function". Doesn't involve derivatives, so it's not Taylor. If you want exactness, I suppose the infinite sums involved have to converge.
 
  • #29
Mark Harder said:
Just a suggestion. Look up the Stone-Weierstrass theorems. The simplest one works for real continuous functions over a real interval. But others work for complex functions, quaternions, etc. For the real case, "every continuous function defined on a closed interval[a, b] can be uniformly approximated as closely as desired by a polynomial function". Doesn't involve derivatives, so it's not Taylor. If you want exactness, I suppose the infinite sums involved have to converge.
This argument needs amendment since ##x## has an unbounded spectrum, while Weierstrass only works for functions of bounded operators.
 
  • #30
A. Neumaier said:
In more explicit terms: Once you know that ##[p,f(x)]=-i\hbar f'(x)## for some function ##f## of an operator vector ##x## with commuting components you get it for ##g(x):=f(x)^{-1}## in place of ##f(x)##, too. By linear combination if you start with ##f(x)=x^n## with a multiexponent ##n## (where it follows from the definition by induction) you get it first for all polynomials, then for their inverses, then for all rational functions. Then for limits of rational functions, for integrals, and Cauchy's integral theorem gives it for all functions analytic on the joint spectrum of ##x## (with exception of any singularities). This is what you need to handle the Coulomb potential.

Its easy to show it for rational functions. But I'm afraid I lack the knowledge for the rest. Can you provide any reference on this?

A. Neumaier said:
Actually, one can do f=1/rf=1/r in an direct way by finding the implicit algebraic equation it satisfies in terms of xx, and proceeding similar to the case of the inverse of a polynomial. This is enjoyable, too, and one doesn't need limits. This works for any algebraic function, but not for exponentials, etc..

I don't quite understand. What algebraic equation do you mean?
 
  • #31
Shyan said:
Can you provide any reference on this?
Not directly. Taking a limit in an operator equation (e.g., to go from polynomials to power series) needs an appropriate topology on operators; these are discussed in books on functional analysis; similarly for the use of Cauchy's theorem (which is complex analysis applied to operators). I recommend that you read the first volume (and the beginning of the second) of the Math Physics book series by Reed and Simon.
Shyan said:
What algebraic equation do you mean?
##f(x)^2(x_1^2+x_2^2+x_3^2)=1##. Apply the commutator with ##p_i## on both sides using the product rule, and collect terms.

By the way, due to limitations of the current copying software here on PO, when you copy part of a post containing equations you often need (as in your previous post) to edit the formulas to make them come out correctly.
 
  • #32
By the functional calculus, ##f(x)## is the operator that multiplies by ##f(x)##. You can then do the calculation directly:
##[p,f]\Psi = -i (f\Psi)^\prime + i f\Psi^\prime = -i f^\prime\Psi - i f \Psi^\prime + i f \Psi^\prime = -i f^\prime\Psi##
 
  • Like
Likes ShayanJ
  • #33
Ju
Shyan said:
I want to calculate the commutator ##{\Large [p_i,\frac{x_j}{r}]}## but I have no idea how I should work with the operator ##{\Large\frac{x_j}{r} }##.
Is it ## x_j \frac 1 r ## or ## \frac 1 r x_j ##? Or these two are equal?
How can I calculate ##{\Large [p_i,\frac 1 r]}##?
Thanks
Just curious, what are you going to do with this?
 
  • #34
rubi said:
By the functional calculus, ##f(x)## is the operator that multiplies by ##f(x)##. You can then do the calculation directly:
##[p,f]\Psi = -i (f\Psi)^\prime + i f\Psi^\prime = -i f^\prime\Psi - i f \Psi^\prime + i f \Psi^\prime = -i f^\prime\Psi##
nice and simple, thanks!
 
  • #35
bob012345 said:
Ju

Just curious, what are you going to do with this?

I'm trying to calculate the commutator of the Hydrogen Hamiltonian with the LRL vector.
 

Similar threads

Replies
1
Views
2K
Replies
9
Views
1K
Replies
10
Views
2K
  • Quantum Physics
Replies
1
Views
605
Replies
9
Views
538
Replies
4
Views
1K
  • Quantum Physics
Replies
5
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
1K
Replies
3
Views
1K
  • Electromagnetism
Replies
2
Views
866
Back
Top