1. The problem statement, all variables and given/known data A glider is tugged by an airplane at 81 m/s when it is released. If the original speed was along the horizontal and the glider is now under a constant acceleration of 2.4 m/s2 at 1.1° below the horizontal due to air drag, how long will it take to reach the ground 5.7 km below? a. 250,000 s b. 500s c. 4.8 s d. 2.2s 2. Relevant equations v0x = v0*cos(theta) v0y = v0*sin(theta) vy = v0y + at x = x0 + v0x*t y = y0 + v0y*t + 1/2*at^2 vy^2 = v0^2 +2a(delta y) 3. The attempt at a solution 1. The problem statement, all variables and given/known data 2. Relevant equations 3. The attempt at a solution No idea on how to start.