- Problem Statement
- A projectile is launched at ground level with an initial speed of 46.0 m/s at an angle of 25.0° above the horizontal. It strikes a target above the ground 2.90 seconds later. What are the x and y distances from where the projectile was launched to where it lands?
- Relevant Equations
Horizontal Motion(ax = 0)
x = x0 + Vx t
Vx = V0x = Vx = velocity is a constant.
Vertical Motion(assuming positive is up ay = −g = −9.80m/s squared)
y = y0 +1/2 (v0y + vy)t
vy = v0y − gt
y = y0 + v0yt −1/2gt2
This based on velocity being constant. Define x0 and y0 to be zero and solve for the desired quantities.