Why is the Jacobian for polar coordinates sometimes negative?

Click For Summary
The discussion centers on the Jacobian determinant in polar coordinates, specifically addressing why it can be negative in certain transformations. It clarifies that while the Jacobian \( J \) can indeed be negative, the absolute value is used in integration formulas. The confusion arises from the different forms of transformations between Cartesian and polar coordinates, which affect the sign of the Jacobian. The importance of the order of variables in transformations is emphasized, as switching them alters the determinant's sign. Ultimately, the correct interpretation is that the Jacobian's absolute value is what matters for integration purposes.
laser
Messages
104
Reaction score
17
Homework Statement
description
Relevant Equations
description
723883159279632384.png


Proving this geometrically [1] gives ##J = r.##

Why is the ##-r## one wrong? Why is ##(x, y) \rightarrow (\theta, r)## is different from ##(x, y) \rightarrow (r, \theta)##? Edit: In Paul's Notes [2] it seems like ##J## is always positive, but online says it can be negative...

[1] The first answer on https://math.stackexchange.com/questions/1656814/how-to-prove-dxdy-r-dr-d-theta
[2] https://tutorial.math.lamar.edu/Classes/CalcIII/ChangeOfVariables.aspx
 
Physics news on Phys.org
I don't think I can delete this post - but I think I've figured out the issue. ##J## can be positive or negative but in the integration formula we take the absolute value of ##J##. My bad for reading it wrong!
 
If you switch a row or column of a determinant, the sign of the determinant changes. Regarding the text in the image you posted, in both cases the transformations are from polar to Cartesian rather than from Cartesian to polar as written in that text. The relevant equations are ##x = r\cos(\theta)## and ##y = r\sin(\theta)##.

The equations for converting from Cartesian to polar are different.
 
laser said:
##J## can be positive or negative but in the integration formula we take the absolute value of ##J##. My bad for reading it wrong!
How is your post related to integration? You didn't mention anything of the sort in your earlier post.
 
Mark44 said:
How is your post related to integration? You didn't mention anything of the sort in your earlier post.

Aside from in the title, which contains "dA = rdrdtheta".

@laser: The change of varaible formula in area integrals is derived from <br /> \mathbf{n}\,dS = \frac{\partial \mathbf{r}}{\partial u} \times \frac{\partial \mathbf{r}}{\partial v}\,du\,dv so that <br /> dS = \left\| \frac{\partial \mathbf{r}}{\partial u} \times \frac{\partial \mathbf{r}}{\partial v} \right\|\,du\,dv. Swapping the order of u and v changes the direction of \mathbf{n} but does not change dS. We treat 2D as being in the plane z = 0, so that the cross-product reduces to <br /> \left(\frac{\partial x}{\partial u} \frac{\partial y}{\partial v} - \frac{\partial x}{\partial v} <br /> \frac{\partial y}{\partial u} \right)\mathbf{e}_z and taking the norm of this gives <br /> \left| \frac{\partial x}{\partial u} \frac{\partial y}{\partial v} - \frac{\partial x}{\partial v} <br /> \frac{\partial y}{\partial u} \right|.
 
  • Like
Likes Orodruin and laser
Just to point out that this change of coordinates is only valid locally, i.e., there's no global change of coordinates in this case.
 
First, I tried to show that ##f_n## converges uniformly on ##[0,2\pi]##, which is true since ##f_n \rightarrow 0## for ##n \rightarrow \infty## and ##\sigma_n=\mathrm{sup}\left| \frac{\sin\left(\frac{n^2}{n+\frac 15}x\right)}{n^{x^2-3x+3}} \right| \leq \frac{1}{|n^{x^2-3x+3}|} \leq \frac{1}{n^{\frac 34}}\rightarrow 0##. I can't use neither Leibnitz's test nor Abel's test. For Dirichlet's test I would need to show, that ##\sin\left(\frac{n^2}{n+\frac 15}x \right)## has partialy bounded sums...