- #1
voxed
- 1
- 0
Homework Statement
An accelerometer measures the acceleration of an object every 0.050s.
The initial velocity is 0 m/s.
The initial distance is 0 m/s.
The accelerations at the times (in m/s^2):
a(0.000) = 5
a(0.050) = 7
a(0.100) = 10
a(0.150) = 15
a(0.200) = 15
a(0.250) = 15
a(0.300) = 5
a(0.350)= -10
a(0.400)= -10
If possible, estimate the distance traveled by the object from t=0s to t=0.400s.
Homework Equations
[tex]
v_{ave} = \Delta x / \Delta t
[/tex]
[tex]
a_{ave} = \Delta v / \Delta t
[/tex]
[tex]
x = x_0 + v_0 t + (1/2) a t^2
[/tex]
[tex]
v^2 = v_0^2 + 2 a \Delta x
[/tex]
The Attempt at a Solution
I tried to find the slopes of the a(t) graph by having v(.025)=(7-5)/(.05-0) and v(.075)=(10-7)/(.1-.05). Then I tried to find the slope of the v(t) graph by x(.05)=(v(.075)-v(.025))/(.075-.025)=400m which is, of course, extremely high for the times and accelerations.
Another approach I thought of was to add up all of the recorded accelerations, divide them by the number of accelerations found, and use this average acceleration in
[tex]
x = x_0 + v_0 t + (1/2) a t^2
[/tex]
as
[tex]
x = 0 + 0 + (1/2) (average acceleration) (total time)^2
[/tex]
Are these correct approaches? Are there any other ways to do this that are more accurate?
Thanks.