Differentiating the complex scalar field

In summary, the conversation revolves around the calculation of partial derivatives in Lagrange's equations for a free scalar field. The main focus is on the differentiation with respect to the complex field and its components, and how to correctly apply the chain rule. Through various explanations and links, it is determined that the correct way to calculate these derivatives is by treating the real and imaginary components separately, using the chain rule, and taking into account the symmetry of the complex field. This resolves the issue of getting incorrect factors in the calculations.
  • #36
A. Neumaier said:
You should realize that the same word may have different meanings in different contexts, being generalized by mathematicians if they can give it a more general interpretation that still fits the formal rules.
Unless of course that word happens to be "observable", right? :smile:

Edit: You're obviously going to counter by saying that you're talking about a generalization while I was talking about a restriction. This means that you aren't contradicting yourself, but I still find it funny that you're so willing to embrace a redefinition of the term "independent" that assigns it to a pair of variables that have "I depend on that guy" tatooed on their foreheads, and at the same time find a restriction of the term "observable" so appalling.



I think I have a pretty good idea about how this Wirtinger stuff works now. This is a summary: Suppose that x,y,z,w are variables that represent complex numbers. In this post I will call any piece of additional information about the values of those variables a constraint. The equalities

[tex]z=x+iy[/tex]

[tex]w=x-iy[/tex]

are constraints. This pair of equalities is equivalent to

[tex]x=\frac{z+w}{2}[/tex]

[tex]y=\frac{z-w}{2i}[/tex]

These constraints implictly define four maps from ℂ2 into ℂ:

[tex](x,y)\mapsto z[/tex]

[tex](x,y)\mapsto w[/tex]

[tex](z,w)\mapsto x[/tex]

[tex](z,w)\mapsto y[/tex]

Now we would like to impose one more constraint, [itex]x\in\mathbb R[/itex]. This is of course equivalent to [itex]w=z*[/itex]. When we do, the maps that are implicitly defined by our constraints change:

[tex](x,y)\mapsto z\qquad :\mathbb R^2\rightarrow\mathbb C[/tex]

[tex](x,y)\mapsto z^*\qquad :\mathbb R^2\rightarrow\mathbb C[/tex]

[tex](z,z^*)\mapsto x\qquad :\{(z,w)\in\mathbb C^2|w=z^*\}\rightarrow\mathbb R[/tex]

[tex](z,z^*)\mapsto y\qquad :\{(z,w)\in\mathbb C^2|w=z^*\}\rightarrow\mathbb R[/tex]

Let's call them u,v,F,G respectively. The partial derivatives of u and v are clearly well-defined, and I don't think it's too horrible to write them as

[tex]\frac{\partial z}{\partial x}=1,\ \frac{\partial z}{\partial y}=i,\ \frac{\partial z^*}

{\partial x}=1,\ \frac{\partial z^*}{\partial y}=-i[/tex]

The definition of partial derivative fails miserably for any function [itex]H:\{(z,w)\in\mathbb C^2|w=z^*\}\rightarrow S[/itex], where S is a subset of ℂ, and this of course includes F and G. The "solution" to this "problem" is apparently to define

[tex]\frac{\partial H(z,z^*)}{\partial z}=\frac{1}{2}\left(\frac{\partial}{\partial x}+i\frac{\partial}{\partial y}\right)\Big((x,y)\mapsto H(u(x,y),v(x,y))\Big)[/tex]

and similarly for the other partial derivative. This definition is motivated by the fact that if the domain of H had been ℂ2, so that the usual definition of partial derivative had worked, [itex]D_1H(u(x,y),v(x,y))[/itex] would have been equal to the right-hand side of the equality above for all [itex]x,y\in\mathbb R[/itex].

So the weird definition of the partial derivatives of a function that's only defined on pairs of the form (z,z*) as a result of the constraint [itex]x,y\in\mathbb R[/itex], is equivalent to just waiting until after we have taken the partial derivatives before we impose that constraint.

What I still don't get about all of this is why we would prefer to make a bunch of weird redefinitions of standard notation and terminology in order to make each step of a nonsensical calculation correct, instead of just saying "hey, let's compute the partial derivative first, and then set w=z*".

Edit: What's even harder to understand is why we would want to describe this result as "z and z* are independent".
 
Last edited:
Physics news on Phys.org
  • #37
A. Neumaier said:
If you don't understand this formula it is only you who is confused.

[ citation needed ]

I referenced 3 books, including 2 you pointed me to. Are you arguing with Ahlfors? Cough up some paper reference where they do the derivation as a limit.


.
 
Last edited:
  • #38
I think A. Neumaier's discussion was clear, and it has helped me to understand this issue much better than I did.

For me, the key point is that partial derivatives with respect a particular variable are well-defined, even if other variables are functions of them.

For example, suppose I have a function g(x,y), where x and y are cartesian coordinates on a plane. The meaning of the partial derivative [itex]\partial g(x,y)/\partial x[/itex] is clear.

Now suppose I am interested in the value of g(x,y) along a curve y=f(x). This is given by g(x,f(x)). But, the partial derivative with respect to x is still well-defined, even though y is no longer "independent" of x.

To be clear, we should write the partial derivative with respect to x in this situation as

[tex]{\partial g(x,y)\over\partial x}\bigg|_{y=f(x)}.[/itex]

Complex derivatives are of this nature, it seems to me. We declare z and z* to be "independent" for purposes of taking partial derivatives, even though we are later going to take z* to be a particular function of z (namely, the complex conjugate).
 
  • #39
Sankaku said:
[ citation needed ]
I referenced 3 books, including 2 you pointed me to. Are you arguing with Ahlfors? Cough up some paper reference where they do the derivation as a limit.

I am not arguing with Ahlfors. I stated a formula which is valid in the Wirtinger calculus, no matter whether or not it is in the book by Ahlfors. One doesn't need a book to see that the limit formula is correct. It follows easily from the other definition. And it can serve as an alternative definition since one can derive from it the formula Ahlfors may have used as definition. (I don't have his book.)
 
  • #40
Avodyne said:
For example, suppose I have a function g(x,y), where x and y are cartesian coordinates on a plane. The meaning of the partial derivative [itex]\partial g(x,y)/\partial x[/itex] is clear.

Now suppose I am interested in the value of g(x,y) along a curve y=f(x). This is given by g(x,f(x)). But, the partial derivative with respect to x is still well-defined, even though y is no longer "independent" of x.

To be clear, we should write the partial derivative with respect to x in this situation as

[tex]{\partial g(x,y)\over\partial x}\bigg|_{y=f(x)}.[/itex]
I think this notation and terminology is very misleading. The worst part is the notation at the end, but let's start at the beginning. g(x,y) isn't a function. That expression represents a member of the range of the function g. If we write g:ℝ2→ℝ, there's no need to mention coordinates.

If we are only interested in the values of g at points in its domain of the form (x,f(x)), we can consider the restriction of g to the set of such points, but the partial derivatives of that function are undefined. What we need to do here is to define the curve C by C(x)=(x,f(x)) for all x, and to consider the ordinary derivative of g°C:ℝ→ℝ. I wouldn't describe the fact that what we're really interested in is an ordinary derivative of a different function than the one we started with, as "the partial derivative with respect to x is still well-defined".

Now let's talk about the notation at the end. If [itex]h:\mathbb R^2\rightarrow\mathbb R[/itex] is differentiable, then the partial derivative with respect to the first variable is the function [itex]D_1h:\mathbb R^2\rightarrow\mathbb R[/itex] defined by

[tex]D_1h(x,y)=\lim_{h\rightarrow 0}\frac{h(x+h,y)-h(x,y)}{h}[/tex]

for all [itex](x,y)\in\mathbb R^2[/itex]. [itex]\partial h/\partial x[/itex] is just an alternative notation for [itex]D_1h[/itex], motivated by the fact that we often use the symbol x as the first variable. The expression

[tex]\frac{\partial h(x,y)}{\partial x}[/tex]

just means "the value of the function [itex]D_1h[/itex] at (x,y)". So

[tex]\frac{\partial f(x,g(x))}{\partial x}[/tex]

can only mean [itex]D_1f(x,g(x))[/itex], which is equal to

[tex]\lim_{h\rightarrow 0}\frac{g(x+h,f(x))-g(x,f(x))}{h},[/tex]

not

[tex]\lim_{h\rightarrow 0}\frac{g(x+h,f(x+h))-g(x,f(x))}{h}=(f\circ C)'(x)[/tex]

Since [itex]\partial/\partial x[/itex] by definition denotes partial differentiation with respect to the first variable (i.e. exactly the same thing as [itex]D_1[/itex]), the expression you used,

[tex]{\partial g(x,y)\over\partial x}\bigg|_{y=f(x)}[/tex]

should therefore be the same thing as

[tex]D_1g(x,y)\big|_{y=f(x)}[/tex]

and I can only interpret that as "what you get when you replace y with f(x) in the expression [itex]D_1g(x,y)[/itex]", and this is [itex]D_1g(x,f(x))[/itex], which is equal to the first of the two limits above, not the second.
 
  • #41
Fredrik said:
You're obviously going to counter by saying that you're talking about a generalization while I was talking about a restriction. This means that you aren't contradicting yourself, but I still find it funny that you're so willing to embrace a redefinition of the term "independent"

My interest in these discussions here on PF is to explain the actual usage of concepts in theoretical physics. One cannot change these traditions, but one can understand them and become confident in their correct use.

In this thread, I was simply explaining in which sense the existing, well-established traditions about df(z^*,z)/dz and ''treating z and z^* as independent'' are fully rigorous and make perfect sense, at least to me.

That you don't like this tradition is a different matter about which I can't argue.
 
  • #42
Suppose we want to solve the equations of motion defined by this Lagrangian.

[tex]
L(\dot{x},\dot{y},x,y) = \frac{1}{2}(\dot{x}^2 + \dot{y}^2) - \frac{C}{2}(x^2 + y^2)
[/tex]

The way 1:

[tex]
0 \;=\; D_t \frac{\partial L}{\partial\dot{x}} - \frac{\partial L}{\partial x} \;=\; \ddot{x} + Cx
[/tex]
[tex]
0 \;=\; D_t \frac{\partial L}{\partial\dot{y}} - \frac{\partial L}{\partial y} \;=\; \ddot{y} + Cy
[/tex]

The way 2:

First we denote [itex]z = x+iy[/itex] and [itex]z^* = x - iy[/itex], and redefine the Lagrangian

[tex]
L = \frac{1}{2}\dot{z}^* \dot{z} - \frac{C}{2} z^* z
[/tex]

(Of course not writing [itex]L(\dot{z},\dot{z}^*,z,z^*)[/itex] explicitly.) Then we assume that [itex]z[/itex] and [itex]z^*[/itex] are independent and compute

[tex]
0 = D_t \frac{\partial L}{\partial \dot{z}^*} - \frac{\partial L}{\partial z^*} = \frac{1}{2}\big(\ddot{z} + Cz\big)
[/tex]

My question is that why would you use the "way 2"? What do you achieve with it? Is it surely worth all the confusion it will inevitably generate? You could have also obtained the same result by the way 1.
 
Last edited:
  • #44
jostpuur said:
Suppose we want to solve the equations of motion defined by this Lagrangian.

[tex]
L(\dot{x},\dot{y},x,y) = \frac{1}{2}(\dot{x}^2 + \dot{y}^2) - \frac{C}{2}(x^2 + y^2)
[/tex]

My question is that why would you use the "way 2"? What do you achieve with it? Is it surely worth all the confusion it will inevitably generate? You could have also obtained the same result by the way 1.

If the Lagrangian is given in your form, there is no reason to perform the transformation.

But suppose you have a problem where your Hamiltonian is given in the form of an anharmonic oscillator
[tex]H(z^*,z)= \omega z^*z + g (z^*z)^2 + g' (z^4+(z^*)^4),[/tex]
say. Then you want to write your dynamics directly in terms of z,
[tex]dz/dt = i dH/dz^*(z^*,z)=\omega z + 2g z(z^*z)^2 + 4g' (z^*)^3,[/tex]
rather than first have to convert it to real and imaginary part, and using the real Hamiltonian equations.

Note that in electrical circuits, say, the variables are naturally given as complex quantities, and the above form is far more natural than the one in terms of real quantities.
 
  • #45
A. Neumaier said:
Sankaku said:
A. Neumaier said:
If H is an analytic function of z^* and z then
[tex] dH(z^*,z)/dz^*=lim_{h\to 0} (H(z^*+h,z)-H(z^*,z))/h[/tex]
makes perfect sense and gives the right result.
I am sorry, I can only say again that z cannot be fixed while you vary [tex]\bar{z}[/tex]. If you can't see the circular logic in your statement, there is nothing I can do.
In case you haven't seen it: I am fixing both z and z^* and vary a _new_ variable h.
There is nothing circular in my argument; it has a standard, perfectly well-defined, rigorous interpretation.

Neumaier, in the beginning you said that H would be an analytic function of z^* and z, which sounds suspicious, because if H is an analytic function of z, then it is not an analytic function of z^*. It could be that this distracted Sankaku. But I see that what you mean makes sense.

Avodyne said:
For me, the key point is that partial derivatives with respect a particular variable are well-defined, even if other variables are functions of them.

I see this now too.

Since this has been a confusing thread, it won't hurt if I iterate this a little bit for others to see more explicitly:

If

[tex]
f:\mathbb{C}^2\to\mathbb{C},\quad (z_1,z_2)\mapsto f(z_1,z_2)
[/tex]

is a function such that it is complex analytic with respect to the both variables separately, then the following partial derivative functions exists

[tex]
(z_1,z_2) \mapsto (\partial_1 f)(z_1,z_2),\quad (z_1,z_2)\mapsto (\partial_2 f)(z_1,z_2)
[/tex]

and it makes sense to use the following notation:

[tex]
\frac{\partial f(z,z^*)}{\partial z} := (\partial_1 f)(z,z^*),\quad \frac{\partial f(z,z^*)}{\partial z^*} := (\partial_2 f)(z,z^*)
[/tex]

Sankaku, Frederik, all clear?

My final comment on this is that it's amazing how physicists succeeded in preventing me from understanding this earlier. ;(
 
  • #46
A. Neumaier said:
Note that in electrical circuits, say, the variables are naturally given as complex quantities, and the above form is far more natural than the one in terms of real quantities.

Everything is real in classical EM unless something is specifically somehow interpreted as complex. Aren't the complex numbers in electrical circuits only used as a computational trick, because people don't want to deal with formulas

[tex]
\sin(A + B) = \sin(A)\cos(B) + \cos(A)\sin(B)
[/tex]
[tex]
\cos(A + B) = \cos(A)\cos(B) - \sin(A)\sin(B)
[/tex]

but prefer

[tex]
e^{i(A + B)} = e^{iA} e^{iB}
[/tex]

instead?
 
  • #47
jostpuur said:
Neumaier, in the beginning you said that H would be an analytic function of z^* and z, which sounds suspicious, because if H is an analytic function of z, then it is not an analytic function of z^*. It could be that this distracted Sankaku. But I see that what you mean makes sense.

Of course. An analytic function in z is as different from an analytic function in z and z^*
as a real function of x and y is different from a real function of x.


jostpuur said:
If
[tex]
f:\mathbb{C}^2\to\mathbb{C},\quad (z_1,z_2)\mapsto f(z_1,z_2)
[/tex]
is a function such that it is complex analytic with respect to the both variables separately, then the following partial derivative functions exists
[tex]
(z_1,z_2) \mapsto (\partial_1 f)(z_1,z_2),\quad (z_1,z_2)\mapsto (\partial_2 f)(z_1,z_2)
[/tex]
and it makes sense to use the following notation:
[tex]
\frac{\partial f(z,z^*)}{\partial z} := (\partial_1 f)(z,z^*),\quad \frac{\partial f(z,z^*)}{\partial z^*} := (\partial_2 f)(z,z^*)
[/tex]

Yes. And in this case one says that f(z,z^*) is an analytic function of z and z^*.

Note that given an analytic function of z and z^* in the form of a nonanalytic function of z (e.g., f=Re z), one can find out what f(z_1,z_2) must be: The series expansion in powers of z and z^* is well-defined and unique. Replacing in this expansion z by z_1 and z^* by z_2 gives the expansion of f(z_1,z_2).
 
  • #48
Fredrik said:
"z and z* are independent variables" should mean that we use the notation (z,z*) for points in the domain of the function we're dealing with. This is of course just as trivial, but if * denotes complex conjugation, the domain of the function would have to be (a subset of) the subset of ℂ2 that consists of pairs of the form (z,z*), and now we have a problem.

Fredrik said:
The definition of partial derivative fails miserably for any function [itex]H:\{(z,w)\in\mathbb C^2|w=z^*\}\rightarrow S[/itex], where S is a subset of ℂ, and this of course includes F and G.

These domains emerged from your attempts to guess the meaning for vague statements, but IMO you should forget them now, because they turned out not to be relevant for sensible interpretations of these initially vague statements.
 
  • #49
jostpuur said:
Aren't the complex numbers in electrical circuits only used as a computational trick, because people don't want to deal with formulas

[tex]
\sin(A + B) = \sin(A)\cos(B) + \cos(A)\sin(B)
[/tex]
[tex]
\cos(A + B) = \cos(A)\cos(B) - \sin(A)\sin(B)
[/tex]

but prefer

[tex]
e^{i(A + B)} = e^{iA} e^{iB}
[/tex]

instead?

I agree with your statement if you drop the ''only'', which isn't justified for a trick that improves things to an extent that it is virtually everywhere used where physicists work with periodic terms. The latter is far more natural than the former, much easier to remember, much easier to use, and in every respect better behaved.

It is the natural expression of periodicity, as can be seen everywhere: from the Fourier transform or from the solution of linear differential equations with constant coefficients in terms of an eigenvalue problem, from the way the Schroedinger equation is treated. And of course also from the way, linear electrical circuits are analyzed. Once your circuit has more than very few elements, it becomes extremely messy to work with trigonometric functions.

All mathematics consists of tricks to make reasoning shorter. We invent the concept of a prime because it is a useful trick not to have to say each time ''number that is not divisible by any other number apart from one itself'', the decimal notation to be able to say 123 in place of ''one times 100 plus two times ten plus three'', etc. We invent the concept of a phase space vector to be able to abbreviate by the letter x a long list of coordinates. etc. etc..

Mathematics progresses by finding concepts that reduce the labor of precise reasoning to an extent that even very complex matters look comprehensible.
 
  • #50
jostpuur said:
My question is that why would you use the "way 2"? What do you achieve with it? Is it surely worth all the confusion it will inevitably generate? You could have also obtained the same result by the way 1.
My problem is not so much with "way 2", but with the way it's presented in physics books. If they had said something like

In this Lagrangian, the symbol * doesn't denote complex conjugation, and z* is just another variable. We determine the equations of motion for these two functions, and find a) that they're exactly the same, and b) that the complex conjugate of any solution is a solution. This means that if we set z* equal to the complex conjugate of z after we have determined the equation satisfied by both, we obtain a theory of a single complex-valued function instead of a theory of two.​

I would have been OK with it.


jostpuur said:
Neumaier, in the beginning you said that H would be an analytic function of z^* and z, which sounds suspicious, because if H is an analytic function of z, then it is not an analytic function of z^*. It could be that this distracted Sankaku. But I see that what you mean makes sense.
It definitely distracted me. I didn't even look at the limit right away. But I agree that the limit is well-defined for all values of z, assuming that H is defined and analytic in an open set that contains the set of pairs of the form (z*,z). This is just the standard (not Wirtinger) definition of partial derivative.

With that in mind, the definition

[tex]\frac{\partial H(z^*,z)}{\partial z}=\frac{1}{2}\left(\frac{\partial}{\partial x}+i\frac{\partial}{\partial y}\right)\Big((x,y)\mapsto H(x-iy,x+iy)\Big)[/tex]

seems strange and unnecessary. This is the sort of stuff that the Wikipedia article put into my head. Maybe it's useful for something, but I don't think we need it here.

jostpuur said:
These domains emerged from your attempts to guess the meaning for vague statements, but IMO you should forget them now, because they turned out not to be relevant for sensible interpretations of these initially vague statements.
I agree that we have no interest in any functions with that domain. All we need to do is to wait until after we have found the partial derivative (which is another function from ℂ2 into ℂ) until we set z*=(complex conjugate of z).

jostpuur said:
My final comment on this is that it's amazing how physicists succeeded in preventing me from understanding this earlier. ;(
I have felt that way many times. I could totally understand the frustration you displayed in #6. I still get angry when I think about how tensors were explained to me in 1994.
 
  • #51
Fredrik said:
My problem is not so much with "way 2", but with the way it's presented in physics books. If they had said something like

In this Lagrangian, the symbol * doesn't denote complex conjugation, and z* is just another variable. We determine the equations of motion for these two functions, and find a) that they're exactly the same, and b) that the complex conjugate of any solution is a solution. This means that if we set z* equal to the complex conjugate of z after we have determined the equation satisfied by both, we obtain a theory of a single complex-valued function instead of a theory of two.​

I would have been OK with it.

They don't say this because that's not what's done.

Suppose z=x+iy is a complex field, and z*=x-iy is its complex conjugate. We can express the lagrangian as a function of x and y, or as a function of z and z*. I will call the first function R and the second function C. (R is to remind us of "real" and C of "complex".) R is a function from ℝ2 into ℝ, and C is a function from ℂ2 into ℂ. These functions are related by

R(x,y)=C(x+iy,x-iy).

From this, it follows immediately (using the chain rule) that the derivatives of these functions are related by (using your non-standard notation)

D1R = (D1 + D2)C

D2R = (1/i)(D1 - D2)C.

We can now solve for D1C and D2C, with the result

D1C = (1/2)(D1 + iD2)R

D2C = (1/2)(D1 - iD2)R.

(These are the formulas that you call "strange and unnecessary".)

The equations of motion are

D1R = 0

D2R = 0.

Using the "strange and unnecessary" formulas, we find that

D1C = 0

D2C = 0.

Now it so happens that it is generally easier to compute D1C and D2C than it is to compute D1R and D2R. One reason for this is that C has the property that C(x+iy,x-iy) must be real, and this implies that D2C=(D1C)* when evaluated at (x+iy,x-iy). Thus we only need to compute D1C.

Note that it is never necessary to say that "the symbol * doesn't denote complex conjugation".
 
Last edited:
  • #52
Avodyne said:
Note that it is never necessary to say that "the symbol * doesn't denote complex conjugation".
OK, you're right. I have no objections to anything you did in this post.

Avodyne said:
your non-standard notation
I think all of these notations are standard, but perhaps I picked the least popular one:

[tex]D_1f(x,y)=\partial_1 f(x,y)=f_{,1}(x,y)=\frac{\partial f(x,y)}{\partial x}=\frac{\partial}{\partial x}f(x,y)=\frac{\partial}{\partial x_1}\bigg|_{(x,y)}f(x_1,x_2)[/tex]

Edit: I think you actually made the best post in the entire thread. It explains a lot, without any weird terminology or strange definitions. One of the things you proved is that there's no need to take the formulas I called "strange and unnecessary" as definitions, because the result

[tex]\frac{\partial C(z,z^*)}{\partial z}=\frac{1}{2}\left(\frac{\partial}{\partial x}+i\frac{\partial}{\partial y}\right)C(x+iy,x-iy)[/tex]

follows from the standard definition of a partial derivative, assuming of course that the function C is defined and analytic in an open subset of ℂ2 that includes the point (z,z*).
 
Last edited:
  • #53
Fredrik said:
OK, you're right. I have no objections to anything you did in this post.
!
Fredrik said:
I think you actually made the best post in the entire thread.
! !
Fredrik said:
I think all of these notations are standard, but perhaps I picked the least popular one:
Well, I haven't seen it before (except in Mathematica). But if the arguments were called [itex](x_1,x_2)[/itex], then I would know what [itex]\partial_1[/itex] meant.

Incidentally, the property that C(x+iy,x-iy) must be real is better expressed as the statement that C must be a symmetric function on ℂ2.
 
  • #54
Fredrik said:
OK, you're right. I have no objections to anything you did in this post.I think all of these notations are standard, but perhaps I picked the least popular one:

[tex]D_1f(x,y)=\partial_1 f(x,y)=f_{,1}(x,y)=\frac{\partial f(x,y)}{\partial x}=\frac{\partial}{\partial x}f(x,y)=\frac{\partial}{\partial x_1}\bigg|_{(x,y)}f(x_1,x_2)[/tex]

Edit: I think you actually made the best post in the entire thread. It explains a lot, without any weird terminology or strange definitions. One of the things you proved is that there's no need to take the formulas I called "strange and unnecessary" as definitions, because the result

[tex]\frac{\partial C(z,z^*)}{\partial z}=\frac{1}{2}\left(\frac{\partial}{\partial x}+i\frac{\partial}{\partial y}\right)C(x+iy,x-iy)[/tex]

follows from the standard definition of a partial derivative, assuming of course that the function C is defined and analytic in an open subset of ℂ2 that includes the point (z,z*).

The point of the Wirtinger calculus is that this assumption (which I made for didactical reasons) is not needed when you work in terms of his definitions. Because in general you only have a function f(z) that is continuous in z and continuously differentiable in Re z and I am z (i.e., considered as a function in R^2), such as F(z)=|Re z|^3-|Im z|^3. One cannot apply Avodyne's recipe in that case, but the Wirtinger derivative 9and the second derivatives) are still well-defined.

It is a bit like the difference between a real analytic function and a real differentiable function. To get the latter, you need more careful definitions than to get the former.
 
  • #55
Avodyne said:
Incidentally, the property that C(x+iy,x-iy) must be real is better expressed as the statement that C must be a symmetric function on ℂ2.

But this is not a requirement for the calculus to work.
 
  • #56
Avodyne said:
!

! !
:smile: Do I seen so stubborn and grumpy that I can't give someone else credit for something good? Maybe it's the avatar. :rolleyes:

Avodyne said:
Well, I haven't seen it before (except in Mathematica).
You made me wonder if the book that I picked up the Dif notation from (a little known Swedish book) was using a non-standard notation, but I just checked my copy of the infamous "baby Rudin" (Principles of mathematical analysis), and it's the default notation there too.

I actually like the notation [itex]f_{,\,i}[/itex] a lot, because it makes the chain rule look so nice, e.g.

[tex](f\circ g)_{,i}(x)=f_{,j}(g(x))g_{j,i}(x)[/tex]
 
  • #57
Fredrik said:
I actually like the notation [itex]f_{,\,i}[/itex] a lot, because it makes the chain rule look so nice, e.g.
[tex](f\circ g)_{,i}(x)=f_{,j}(g(x))g_{j,i}(x)[/tex]

The notation without indices (regarding D as a vector with components D_i) looks even nicer: D(f(g(x)) = Df(g(x))Dg(x)
 
  • #58
Arrggh! I somehow managed to screw up some minus signs in my big post above, and now I'm not seeing the "edit" button.

Correct versions are

D1R = (D1 + D2)C

D2R = i(D1 - D2)C.

We can now solve for D1C and D2C, with the result

D1C = (1/2)(D1 - iD2)R

D2C = (1/2)(D1 + iD2)R.

This implies

[tex]{\partial\over\partial z}C(z,z^*)={1\over2}\left({\partial\over\partial x}-i{\partial\over\partial y}\right)C(x+iy,x-iy)[/tex]

which is the same formula given by A.Neumaier in post #17.
 
  • #59
A. Neumaier said:
in general you only have a function f(z) that is continuous in z and continuously differentiable in Re z and I am z (i.e., considered as a function in R^2), such as F(z)=|Re z|^3-|Im z|^3. One cannot apply Avodyne's recipe in that case

I don't see the problem ... Is it because of the absolute-value signs?
 
  • #60
Fredrik said:
:smile: Do I seen so stubborn and grumpy that I can't give someone else credit for something good? Maybe it's the avatar. :rolleyes:
I thought my math would be too "lowbrow" for you, though I did make an effort to specify the domain and range of my functions, and to avoid sloppy language, such as referring to the function R as R(x,y).
 
  • #61
Avodyne said:
I don't see the problem ... Is it because of the absolute-value signs?

Yes. The absolute value is not an analytic function. i.e., you can't expand it into a power series.
 

Similar threads

Replies
5
Views
470
Replies
3
Views
1K
Replies
3
Views
811
Replies
4
Views
907
  • Quantum Physics
Replies
2
Views
746
  • Quantum Physics
Replies
13
Views
797
  • Quantum Physics
Replies
1
Views
845
  • Quantum Physics
Replies
1
Views
630
  • Quantum Physics
Replies
4
Views
898
  • Quantum Physics
Replies
4
Views
1K
Back
Top