Coulomb potential as an operator

Click For Summary

Discussion Overview

The discussion revolves around the calculation of the commutator involving the momentum operator and the Coulomb potential expressed as an operator. Participants explore various mathematical approaches to handle the operator ##{\Large\frac{x_j}{r}}## and the implications of using different forms of functions in the context of quantum mechanics.

Discussion Character

  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant questions how to work with the operator ##{\Large\frac{x_j}{r}}##, asking if it should be treated as ## x_j \frac{1}{r} ## or ## \frac{1}{r} x_j ##.
  • Another participant suggests that the momentum operator acts as a derivative on a function of the operator ##r##.
  • It is noted that since ##x_j## commute and ##r^2=\sum x_j^2##, the operators ##x## and ##r## commute, leading to a method of differentiation to find partial derivatives.
  • Concerns are raised about the applicability of the commutator formula ##[p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k}## for functions that are not polynomials, specifically for ##\frac{1}{r}##.
  • Some participants argue that the commutator can apply to any smooth function, while others challenge this by emphasizing the need for proof and clarification on the nature of ##f(x)##.
  • Discussions include the Taylor expansion of functions and its limitations, particularly regarding the expansion of functions like ##f(x)=(1+x)^{-1}##.
  • Participants explore the implications of commutation relations for operators and how they extend to rational functions and analytic functions.
  • There is a proposal to find a direct method to handle the operator ##f=1/r## without relying on limits or Taylor expansions.

Areas of Agreement / Disagreement

Participants express differing views on the applicability of certain mathematical techniques, particularly regarding the use of Taylor expansions and the nature of functions involved in the commutator calculations. No consensus is reached on the best approach to handle the Coulomb potential as an operator.

Contextual Notes

Limitations include the dependence on definitions of smooth functions and the unresolved nature of the mathematical steps involved in proving the commutation relations for non-polynomial functions.

ShayanJ
Science Advisor
Insights Author
Messages
2,802
Reaction score
605
I want to calculate the commutator ##{\Large [p_i,\frac{x_j}{r}]}## but I have no idea how I should work with the operator ##{\Large\frac{x_j}{r} }##.
Is it ## x_j \frac 1 r ## or ## \frac 1 r x_j ##? Or these two are equal?
How can I calculate ##{\Large [p_i,\frac 1 r]}##?
Thanks
 
Physics news on Phys.org
My first thought is that the momentum operator act as a derivative on a function of the operator r
 
Since the ##x_j## commute and ##r^2=\sum x_j^2##, ##x## and ##r## commute. You can differentiate the relation defining ##r## to find the partial derivatives, and then use ##[p_k,f(x)]=-i\hbar \partial f(x)/\partial x_k##. (Don't use ##i## as index if it appears as ##\sqrt{-1}##, too.)
 
Last edited:
I've been able to prove ## [p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k} ## only for the case of ## f(x) ## being a polynomial in ## x_k ##. But here ## \frac 1 r ## doesn't satisfy this condition. Is there a general proof?
 
Last edited:
Shyan said:
##[p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k}##
Actually the differential operator should be of partial one.
Shyan said:
I've been able to prove ## [p_k,f(x)]=-i\hbar \frac{df(x)}{dx_k} ## only for the case of ## f(x) ## being a polynomial in ## x_k ##. But here ## \frac 1 r ## doesn't satisfy this condition. Is there a general proof?

I have no idea what you're talking about!
The commutator suggested by Neumaier is applicable to any smooth function ##f(x)##, not only polynomials.
 
blue_leaf77 said:
The commutator suggested by Neumaier is applicable to any smooth function
## f(x) ## is a function of positon operator, not a smooth function of a real variable!
 
Who said that ##f(x)## is a function of real variable?
 
blue_leaf77 said:
Who said that ##f(x)## is a function of real variable?
I got the impression that you meant it was trivial since you didn't provide any proof, so I thought that's what you mean. Do you know where can I find a proof?
 
Consider
$$
[p,x^n] = -i\hbar nx^{n-1}
$$
and the Taylor expansion of ##f(x)##.
 
  • #10
blue_leaf77 said:
Consider
$$
[p,x^n] = -i\hbar nx^{n-1}
$$
and the Taylor expansion of ##f(x)##.

A while ago I asked about this and micromass didn't seem to agree with you.(See here!)
Actually that's why I didn't think about using the formula in the first place.
 
  • #11
It seems that Taylor expansion of a function operator cannot be taken lightly, I have actually also addressed exactly the same problem on this regard almost a year ago There I asked if a function ##f(x)=(1+x)^{-1}## can be expanded in power terms, as the same function with ##x## real numbers can only have Taylor expansion for a bounded x ##-1<x<1##.
 
  • #12
blue_leaf77 said:
It seems that Taylor expansion of a function operator cannot be taken lightly, I have actually also addressed exactly the same problem on this regard almost a year ago There I asked if a function ##f(x)=(1+x)^{-1}## can be expanded in power terms, as the same function with ##x## real numbers can only have Taylor expansion for a bounded x ##-1<x<1##.
Is there any proof that doesn't use Taylor expansion of operator functions?
 
  • #13
Shyan said:
Is there any proof that doesn't use Taylor expansion of operator functions?
Well, it starts from Taylor expansion of ordinary functions. But yes, one can extend to very general (complex analytic) functions.

To get started, restrict to the case where an (ordinary, polynomial) operator ##A## has an inverse ##A^{-1}##. I.e., ##A A^{-1} = 1##. Then take the commutator with ##p## on both sides, and use the Leibniz rule on the LHS. What do you find? :oldwink:[And thank you to @A. Neumaier for teaching me this generalized material ages ago.]
 
  • Like
Likes   Reactions: blue_leaf77
  • #14
strangerep said:
To get started, restrict to the case where an (ordinary, polynomial) operator A has an inverse A^{-1}. I.e., A A^{-1} = 1. Then take the commutator with p on both sides, and use the Leibniz rule on the LHS. What do you find? :oldwink:
For any operator B, I find ## [B,A]A^{-1}=A[A^{-1},B] ##! So?
 
  • #15
Shyan said:
For any operator B, I find ## [B,A]A^{-1}=A[A^{-1},B] ##! So?
Take it further...

When ##A = A(x)## (ordinary polynomial), and ##B = p##, and you've already found that ##[p, A(x)] = A'(x)## (modulo factors of ##i## and ##\hbar##, depending on your conventions),... then what is $$[p, A^{-1}(x)] ~=~ ?$$
 
  • #16
strangerep said:
Take it further...

When ##A = A(x)## (ordinary polynomial), and ##B = p##, and you've already found that ##[p, A(x)] = A'(x)## (modulo factors of ##i## and ##\hbar##, depending on your conventions),... then what is $$[p, A^{-1}(x)] ~=~ ?$$
I don't see how ## [p,A]A^{-1}=A[A^{-1},p] ## implies ##[p,A(x)]\propto A'(x)##!
 
  • #17
Shyan said:
I don't see how ## [p,A]A^{-1}=A[A^{-1},p] ## implies ##[p,A(x)]\propto A'(x)##!
You already had that back in post #4. Now use that to find ##[p,A^{-1}] = ?##
 
  • #18
## [p,A(x)^{-1}]=i\hbar A(x)^{-1} A'(x) A^{-1}(x)##
 
  • #19
You keep stopping too soon.

Let ##F(x) := A^{-1}(x)##. From the above, what is ##[p, F(x)] = \; ?##
 
  • #20
Can you be sure that A^-1 commutes with A' ?
 
  • #21
andresB said:
Can you be sure that A^-1 commutes with A' ?
They're both functions of x so I think yes, which gives ## [p,A^{-1}(x)]=i\hbar [A(x)^{-1}]^2 A'(x) ##.

strangerep said:
You keep stopping too soon.

Let ##F(x) := A^{-1}(x)##. From the above, what is ##[p, F(x)] = \; ?##
So I have ## [p,F(x)]=i\hbar [F(x)]^2 A'(x) ##. But what is ##A'(x)## in terms of ##F(x)##?
 
  • #22
andresB said:
Can you be sure that A^-1 commutes with A' ?
Sorry -- I should have specified that ##A^{-1}## is 2-sided inverse operator. But in the cases we're considering, where ##A## is a function of a single operator ##x##, it's a reasonably safe assumption, at least for purposes of physics (with appropriate caveats on the domain of definition).
 
  • #23
Shyan said:
So I have ## [p,F(x)]=i\hbar [F(x)]^2 A'(x) ##. But what is ##A'(x)## in terms of ##F(x)##?
Oh well, I guess we can all be a bit blind sometimes. Certainly, it happens to me.

You have essentially proven that $$[p, A^{-1}(x)] ~=~ \frac{d A^{-1}(x)}{dx} ~.$$ This generalizes the earlier derivative formula to a larger class of functions than mere polynomials.

From this, you can go further, to functions of the form ##f(x) = g(x)/h(x)##, and generalize the derivative formula even further.
 
  • Like
Likes   Reactions: ShayanJ
  • #24
Shyan said:
## [p,A(x)^{-1}]=i\hbar A(x)^{-1} A'(x) A^{-1}(x)##
strangerep said:
You have essentially proven that ##[p, A^{-1}(x)] ~=~ \frac{d A^{-1}(x)}{dx}##. This generalizes the earlier derivative formula to a larger class of functions than mere polynomials.

From this, you can go further, to functions of the form ##f(x) = g(x)/h(x)##, and generalize the derivative formula even further.
In more explicit terms: Once you know that ##[p,f(x)]=-i\hbar f'(x)## for some function ##f## of an operator vector ##x## with commuting components you get it for ##g(x):=f(x)^{-1}## in place of ##f(x)##, too. By linear combination if you start with ##f(x)=x^n## with a multiexponent ##n## (where it follows from the definition by induction) you get it first for all polynomials, then for their inverses, then for all rational functions. Then for limits of rational functions, for integrals, and Cauchy's integral theorem gives it for all functions analytic on the joint spectrum of ##x## (with exception of any singularities). This is what you need to handle the Coulomb potential.
 
Last edited:
  • Like
Likes   Reactions: ShayanJ and blue_leaf77
  • #25
A. Neumaier said:
In more explicit terms: [...]
Yes -- that's where I was heading. But I didn't want to "deprive" Shyan of the enjoyment I experienced (some time ago) of discovering these things by working them out. :biggrin:
 
  • #26
strangerep said:
Yes -- that's where I was heading. But I didn't want to "deprive" Shyan of the enjoyment I experienced (some time ago) of discovering these things by working them out. :biggrin:
I think he'll still enjoy it with my shortcut guide rather than a long interactive guide. There are enough steps to fill in (and small pitfalls) to make it mathematically rigorous.

Actually, one can do ##f=1/r## in an direct way by finding the implicit algebraic equation it satisfies in terms of ##x##, and proceeding similar to the case of the inverse of a polynomial. This is enjoyable, too, and one doesn't need limits. This works for any algebraic function, but not for exponentials, etc..
 
  • #27
I have to wait for 11 months to find a satisfying answer to this problem of reciprocal operator.
 
  • #28
Shyan said:
Is there any proof that doesn't use Taylor expansion of operator functions?

Just a suggestion. Look up the Stone-Weierstrass theorems. The simplest one works for real continuous functions over a real interval. But others work for complex functions, quaternions, etc. For the real case, "every continuous function defined on a closed interval[a, b] can be uniformly approximated as closely as desired by a polynomial function". Doesn't involve derivatives, so it's not Taylor. If you want exactness, I suppose the infinite sums involved have to converge.
 
  • #29
Mark Harder said:
Just a suggestion. Look up the Stone-Weierstrass theorems. The simplest one works for real continuous functions over a real interval. But others work for complex functions, quaternions, etc. For the real case, "every continuous function defined on a closed interval[a, b] can be uniformly approximated as closely as desired by a polynomial function". Doesn't involve derivatives, so it's not Taylor. If you want exactness, I suppose the infinite sums involved have to converge.
This argument needs amendment since ##x## has an unbounded spectrum, while Weierstrass only works for functions of bounded operators.
 
  • #30
A. Neumaier said:
In more explicit terms: Once you know that ##[p,f(x)]=-i\hbar f'(x)## for some function ##f## of an operator vector ##x## with commuting components you get it for ##g(x):=f(x)^{-1}## in place of ##f(x)##, too. By linear combination if you start with ##f(x)=x^n## with a multiexponent ##n## (where it follows from the definition by induction) you get it first for all polynomials, then for their inverses, then for all rational functions. Then for limits of rational functions, for integrals, and Cauchy's integral theorem gives it for all functions analytic on the joint spectrum of ##x## (with exception of any singularities). This is what you need to handle the Coulomb potential.

Its easy to show it for rational functions. But I'm afraid I lack the knowledge for the rest. Can you provide any reference on this?

A. Neumaier said:
Actually, one can do f=1/rf=1/r in an direct way by finding the implicit algebraic equation it satisfies in terms of xx, and proceeding similar to the case of the inverse of a polynomial. This is enjoyable, too, and one doesn't need limits. This works for any algebraic function, but not for exponentials, etc..

I don't quite understand. What algebraic equation do you mean?
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 7 ·
Replies
7
Views
993
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 59 ·
2
Replies
59
Views
6K
  • · Replies 1 ·
Replies
1
Views
1K