- #1
MooCow
- 6
- 0
Problem:
Assume that f has a derivative everywhere. Set g(x)=xf(x). Using the definition of the derivative, show that g has a derivative and that g'(x)=f(x)+xf'(x).
What I know:
I know the definition of the derivative is [f(x+h)-f(x)]/h. I don't know how to plug it in and solve it though. I tried just plugging it straight in like
g'(x)=[x(f(x+h))-xf(x)]/h
I pulled the x out and the rest was [f(x+h)-f(x)]/h (the actual definition) and assumed it was xf'(x). I don't think that's right though.
Any help would be great and appreciated. Thanks in advance:]
Assume that f has a derivative everywhere. Set g(x)=xf(x). Using the definition of the derivative, show that g has a derivative and that g'(x)=f(x)+xf'(x).
What I know:
I know the definition of the derivative is [f(x+h)-f(x)]/h. I don't know how to plug it in and solve it though. I tried just plugging it straight in like
g'(x)=[x(f(x+h))-xf(x)]/h
I pulled the x out and the rest was [f(x+h)-f(x)]/h (the actual definition) and assumed it was xf'(x). I don't think that's right though.
Any help would be great and appreciated. Thanks in advance:]