1. The problem statement, all variables and given/known data The length of a simple pendulum is about 100 cm known to have an accuracy of 1 mm. Its period of oscillation is 2 s determined by measuring the time for 100 oscillations using a clock of 0.1 s resolution. What is the accuracy in the determined value of g? (a) 0.2% (b) 0.5% (c) 0.1% (d) 2% 2. Relevant equations 3. The attempt at a solution →Now, l=100 cm=1 m, Δl=1 mm= 10-3m, t=2 s Δt=0.1 s Then, T=2π✓(l/g) Therefore, g=4π2l/T2 Hence, max error is Δg/g = Δl/l + 2(Δt/t) i.e., max percentage error is (Δg/g)100 = (Δl/l)100 + 2(Δt/t)100 =10.1% Solving this, I'm getting 10.1%, which is obviously not matching with any of the choices. Where am I going wrong? How do I use the number of oscillations given?