(adsbygoogle = window.adsbygoogle || []).push({}); 1. The problem statement, all variables and given/known data

A plane is moving at 600m/s with respect to the ground. According to clocks on the ground, how much time would it take so that the plane's clock is delayed by 2 microseconds?

2. Relevant equations

Lorentz transformations.

3. The attempt at a solution

Let O be a reference frame on the ground and O' be a reference frame on the plane.

v=600m/s. If I'm not wrong, they ask me [tex]t_B-t_A[/tex] such that [tex](t_B-t_A)-(t_A'-t_B')=2 \times 10 ^{-6}s[/tex]. (*)

What I've done so far is [tex]t_B'-t_A'=\gamma \left [ t_B-t_A +\frac{v}{c^2}(x_A-x_B) \right ][/tex], replacing [tex]x_A-x_B[/tex] by [tex]v(t_A-t_B)[/tex], then solving for [tex]t_B-t_A[/tex] in (*), I reach that it's worth exactly [tex]1000000s[/tex]. Or 11 days, 13 hours, 46 minutes and 40 s. It seems too big for me. Do you get a different answer?

**Physics Forums | Science Articles, Homework Help, Discussion**

Dismiss Notice

Join Physics Forums Today!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

# Homework Help: Special relativity, delay of a clock in a plane

**Physics Forums | Science Articles, Homework Help, Discussion**