f(x) = sin(x) for x < or = 0 and

f(x) = x for x > 0

Question was...does f'(x) exist? what is it?

-------------

First I proved the f(x) is continuous at 0 by stating limit as x->0 from left = limit as x->0 from right = f(0).

Then...taking the derivative turned out to be a problem.

the lim[f(x+h)-f(x)]/h was unclear because...if x=0 and you add some h to it, that'd mean you use as your f(x) the x...but point (0, 0) belongs to Sin(x) not to x.

What i did is state that in x-values close to zero, from left and from right, the f'(x) = 1. Basically I ruled out situations like |x| where dy/dx = -1 and 1 for values around x=0.

I assumed that if f'(-0.0001) and f'(0.0001) are equal, the function (who we know is continuous) must be differentiable at that point.

---------------

But my question is...is my logic something that will make my teacher grab his hairs before giving me a nice zero? or a 'wise' way to look the issue?

Also, is there a better, clearer way to solve this?

I always experience uncertainty on what to do and to which expression when a function gets split according to Domain.

Thank you.

~Robokapp