Hi all,(adsbygoogle = window.adsbygoogle || []).push({});

is there a general way of proving that

sqrt(r1) + sqrt(r2) + sqrt(r3) + ... + sqrt(rn) is irrational, given that none of r1, r2, r3, ..., rn is the square of a rational number?

(or is this statement even true in general?)

for the case when n = 2, the proof is quite straight-forward; i think it can be found in most elementary textbooks.

Letting sqrt(a) + sqrt(b) = r, where r is rational, we have

sqrt(a) - sqrt(b) = (a - b) / r = q, where q is rational.

Therefore adding the two equations and halving the result gives

sqrt(a) = 1/2(r + q), which is rational, contradicting our hypothesis.

i tried to extend this proof to the case n = 3, although my proof is quite clumsy and i'm not sure whether it's correct.

however, i am interested to know whether it is true for all n, and if so how it can be proved. thanks for sharing :)

**Physics Forums - The Fusion of Science and Community**

Join Physics Forums Today!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

# One more elementary question, on square roots

Loading...

Similar Threads - more elementary question | Date |
---|---|

B Proof of elementary row matrix operation. | Jun 6, 2017 |

B Sets and functions that gain more structure with context | Jul 7, 2016 |

I Eigen Vectors, Geometric Multiplicities and more... | Apr 10, 2016 |

What is more efficient, autocorrelation or SSA? | May 2, 2015 |

**Physics Forums - The Fusion of Science and Community**