|Sep15-04, 04:18 AM||#1|
Triple Integral in Spherical Coordinates
I have a hemispherical surface of radius R with it's base centred on the origin. We are using the convention:
r is the radius i.e. the magnitude of the position vector of a point: its distance from the origin.
theta is the polar angle
phi is the azimuthal angle
I am asked to calculate the integral of the divergence of a given vector field v over this volume (enclosed by the hemisphere).
So far what I have done is to say that a full sphere would be given by the equation:
r = R
What about a hemisphere? It seems to me that the angle theta must be restricted so that points below the x-y plane are not part of the domain. So what are the allowed values of theta? I can't seem to figure out whether it should be -pi/2 < theta < pi/2, or something else? I need to know this to set my bounds of integration for one of the three integrals. Thanks.
|Sep15-04, 06:36 AM||#2|
For a hemisphere : r goes from 0 to R
theta goes from : 0 to 90°
phi goes from : 0 to 360°
Ofcourse, you know that you need to express the angles in radials
|Similar Threads for: Triple Integral in Spherical Coordinates|
|triple integration w/spherical coordinates||Calculus & Beyond Homework||3|
|Triple integral over a sphere in rectangular coordinates||Calculus & Beyond Homework||5|
|triple integrals in spherical coordinates||Introductory Physics Homework||2|
|How to setup an integral in spherical coordinates for the volume of p = 2 sin O(theta||Introductory Physics Homework||3|
|Setting up a triple integral in cylindrical coordinates?||Introductory Physics Homework||5|