
#1
Oct506, 12:32 PM

P: 218

It's a question I had on a quiz a few minutes ago.
f(x) = sin(x) for x < or = 0 and f(x) = x for x > 0 Question was...does f'(x) exist? what is it?  First I proved the f(x) is continuous at 0 by stating limit as x>0 from left = limit as x>0 from right = f(0). Then...taking the derivative turned out to be a problem. the lim[f(x+h)f(x)]/h was unclear because...if x=0 and you add some h to it, that'd mean you use as your f(x) the x...but point (0, 0) belongs to Sin(x) not to x. What i did is state that in xvalues close to zero, from left and from right, the f'(x) = 1. Basically I ruled out situations like x where dy/dx = 1 and 1 for values around x=0. I assumed that if f'(0.0001) and f'(0.0001) are equal, the function (who we know is continuous) must be differentiable at that point.  But my question is...is my logic something that will make my teacher grab his hairs before giving me a nice zero? or a 'wise' way to look the issue? Also, is there a better, clearer way to solve this? I always experience uncertainty on what to do and to which expression when a function gets split according to Domain. Thank you. ~Robokapp 



#2
Oct506, 12:37 PM

Sci Advisor
HW Helper
PF Gold
P: 12,016

Split your argument in two:
1. Find the righthand derivative 2. Find the lefthand derivative If they are equal, then your function is differentiable at 0. 



#3
Oct506, 12:37 PM

Sci Advisor
HW Helper
P: 9,398

f is trivially differentiable for x>0 and x<0. At x=0, one merely needs to verify if the left and right derivatives exist and agree. They trivially exist and as cos(x) tends to 1 as x tends to zero they agree.



Register to reply 
Related Discussions  
second type perpetual motion machine is existent .  Classical Physics  2 