1. The problem statement, all variables and given/known data Calculate the difference in time after one year between a clock at Earth's surface and a clock on a satellite orbiting at 300 km above the surface 2. Relevant equations T = T0 / (1 - 2gR/c^2)^.5 That is, this: http://hyperphysics.phy-astr.gsu.edu/hbase/relativ/imgrel/gtim3.gif 3. The attempt at a solution I don't understand how to use this equation to get the difference between the clock on the satellite and the clock on the surface. Do I just take the value of T with R = Earth's radius, and again with R = Earth's radius + 300 km, and take the difference?
Hi Seedling! (try using the X^{2} and X_{2} tags just above the Reply box ) The ratio, rather than the difference … T_{0} is the time on a clock "at infinity", and T_{R} is the time on a clock at radius R, so T_{R}/T_{0} is the ratio of their "speeds", and T_{R=h}/T_{R} is the ratio you want. (But remember that this formula only gives you the general relativity time dilation … there'll also be an ordinary Lorentz time dilation, in the opposite direction ).