- #1
bawbag
- 13
- 1
Given a sphere [tex] x^2 + y^2 + z^2 = a^2[/tex] how would I derive the surface area by using surface integrals?
The method I've tried is as follows: [tex]dA = sec\ \gamma \ dxdy[/tex] where gamma is the angle between the tangent plane at dA and the xy plane. [tex]sec \gamma = \frac{|\nabla \varphi|}{\partial \varphi /\partial z}[/tex][tex] = \frac{\sqrt{(2x)^2 + (2y)^2 + (2z)^2}}{2z}[/tex][tex] = \frac{a}{\sqrt{a^2 - x^2 - y^2}}[/tex]
converting to polar coordinates and integrating the expression for [itex]sec \gamma[/itex][tex]\int^{a}_{0}\int^{2 \pi}_{0} \frac{a}{\sqrt{a^2 - r^2}} r \ drd \theta[/tex] and using substitution [itex] u = a^2 - r^2[/itex] yields [tex] -a \pi [2 \sqrt{a^2 - a^2} - 2 \sqrt{a^2 - 0}][/tex][tex] = 2 \pi a^2[/tex] Obviously the area of a sphere is [itex] 4 \pi a^2[/itex] so did I do the calculation wrong, or does this method only find the area of one hemisphere, in which case I should multiply the answer by 2?
Thanks in advance
The method I've tried is as follows: [tex]dA = sec\ \gamma \ dxdy[/tex] where gamma is the angle between the tangent plane at dA and the xy plane. [tex]sec \gamma = \frac{|\nabla \varphi|}{\partial \varphi /\partial z}[/tex][tex] = \frac{\sqrt{(2x)^2 + (2y)^2 + (2z)^2}}{2z}[/tex][tex] = \frac{a}{\sqrt{a^2 - x^2 - y^2}}[/tex]
converting to polar coordinates and integrating the expression for [itex]sec \gamma[/itex][tex]\int^{a}_{0}\int^{2 \pi}_{0} \frac{a}{\sqrt{a^2 - r^2}} r \ drd \theta[/tex] and using substitution [itex] u = a^2 - r^2[/itex] yields [tex] -a \pi [2 \sqrt{a^2 - a^2} - 2 \sqrt{a^2 - 0}][/tex][tex] = 2 \pi a^2[/tex] Obviously the area of a sphere is [itex] 4 \pi a^2[/itex] so did I do the calculation wrong, or does this method only find the area of one hemisphere, in which case I should multiply the answer by 2?
Thanks in advance