Ok heres the problem: A batter hits a ball which leaves the bat 1 meter above the ground at an angle of 65 degrees with an initial velocity of 30 m/s. How far from home plate will the ball land if not caught and ignoring any air resistence? So I tried solving for how much of the velocity is is the x and y components. X= Vcos65 Y= Vsin65 which gave me x=12.68 and y=27.19 Then using the equation X=X(initial) + V(initial x)T + .5A(x)T^2 I plugged in my numbers to get 0=1+12.68t+.5(9.8)T^2. This is where I'm stuck. Am I on the right track or what? Help please. Thanks.