Showing That a Function Does Not Have Two Distinct Roots

  • Context: Undergrad 
  • Thread starter Thread starter Expiring
  • Start date Start date
  • Tags Tags
    Function Roots
Click For Summary

Discussion Overview

The discussion revolves around the proof that the function \(x^3 - 3x + m = 0\) does not have two distinct roots within the interval \(0 \leq x \leq 1\). Participants explore the application of Rolle's Theorem in this context and clarify the conditions under which the proof is valid.

Discussion Character

  • Technical explanation
  • Debate/contested
  • Conceptual clarification

Main Points Raised

  • One participant presents a proof using Rolle's Theorem, asserting that if there are two distinct roots \(a\) and \(b\) in the interval, then there must exist a point \(c\) where \(f'(c) = 0\), leading to a contradiction since \(c\) is found outside the interval.
  • Another participant questions whether the interval \(0 \leq x \leq 1\) was part of the original problem statement, suggesting it may have been an addition by the proof author.
  • The proof author acknowledges the oversight and confirms that the interval was indeed part of the original problem.
  • A later reply indicates that the proof is acceptable with the clarified interval, allowing the assumption that \(0 < a < b < 1\) without loss of generality.

Areas of Agreement / Disagreement

Participants generally agree on the validity of the proof once the interval is clarified, but there is an initial uncertainty regarding the inclusion of the interval in the original problem statement.

Contextual Notes

The discussion highlights the importance of clearly stating problem conditions, as the initial omission led to confusion regarding the proof's applicability. The reliance on Rolle's Theorem is contingent on the assumptions made about the roots and the interval.

Expiring
Messages
4
Reaction score
3
TL;DR
Showing that the function
x^3 - 3x + m = 0
does not have two distinct roots on the interval 0 <= x <= 1 for any value of m using Rolle's Theorem.
I am wondering if someone can look over my proof, and point out any mistakes I might have made.There is no value of m such that
x^3 - 3x + m = 0
has two distinct roots on the interval 0 <= x <= 1.

Proof.

Let f(x) = x^3 - 3x + m. Suppose, to the contrary, that there is a value of m such that f has two distinct roots in 0 <= x <= 1, and suppose that these roots occur within the interval at x=a and x=b. We apply Rolle's Theorem on the interval (a, b). The conditions of Rolle's Theorem are met since
1.) f is a polynomial, so it is continuous and differentiable everywhere, and, as a result, is continuous and differentiable on (a, b),
2.) f(a)=f(b)=0.
Then, according to Rolle's Theorem, there exists some point x=c in the interval 0<a<c<b<1 such that f'(c)=0. Computing f'(c) and finding c we see that c=-1, 1. Since both of these points lie outside the interval (0, 1), they lie outside the interval (a, b). We have reached a contradiction.
 
Last edited:
  • Like
Likes   Reactions: PeroK
Physics news on Phys.org
Expiring said:
TL;DR Summary: Showing that the function
x^3 - 3x + m = 0
does not have two distinct roots for any value of m using Rolle's Theorem.

I am wondering if someone can look over my proof, and point out any mistakes I might have made.There is no value of m such that
x^3 - 3x + m = 0
has two distinct roots on the interval 0 <= x <= 1.
This interval doesn't appear in the original problem description shown in the summary. Is it actually a part of the problem that you neglected to show, or is this something that you added that isn't part of the given problem?
Expiring said:
Proof.

Let f(x) = x^3 - 3x + m. Suppose, to the contrary, that there is a value of m such that f has two distinct roots in 0 <= x <= 1, and suppose that these roots occur within the interval at x=a and x=b. We apply Rolle's Theorem on the interval (a, b). The conditions of Rolle's Theorem are met since
1.) f is a polynomial, so it is continuous and differentiable everywhere, and, as a result, is continuous and differentiable on (a, b),
2.) f(a)=f(b)=0.
Then, according to Rolle's Theorem, there exists some point x=c in the interval 0<a<c<b<1 such that f'(c)=0. Computing f'(c) and finding c we see that c=-1, 1. Since both of these points lie outside the interval (0, 1), they lie outside the interval (a, b). We have reached a contradiction.
 
Mark44 said:
This interval doesn't appear in the original problem description shown in the summary. Is it actually a part of the problem that you neglected to show, or is this something that you added that isn't part of the given problem?
It is part of the problem that I mistakenly left out. I fixed my original post.
 
With that change, your proof looks fine. You can assume, without loss of generality (wlog), that 0 < a < b < 1.
 

Similar threads

  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 22 ·
Replies
22
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 13 ·
Replies
13
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 20 ·
Replies
20
Views
4K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 4 ·
Replies
4
Views
1K