- #1

- 476

- 0

Given that:

y' = -gt + A y' is the vertical velocity

y = -(1/2).g.t^2 +At +B y is the distance above the ground

In order to solve how long takes skydiver to hit ground I did the following:

0 = -(1/2).g.t^2 +(0)t + 4000

4000 = -(1/2).g.t^2

8000 = g.t^2

800 = t^2

t = 28.284 seconds!

Now from here I am unsure how to find the speed at which hits the ground. Help needed sorry for bad formatting!