The Math Challenge - January 2021 features various mathematical problems encompassing topics such as linear programming, calculus, and group theory, with several problems successfully solved by participants. Key discussions include proving the irrationality of numbers like π and e^2, solving differential equations, and demonstrating properties of matrices and integrals. Participants also explored the uniqueness of solutions in differential equations and properties of harmonic numbers. The thread showcases a collaborative environment where complex mathematical concepts are tackled and clarified through peer interaction.
#31
AndreasC
Gold Member
555
317
If I wanted to pretend I knew how to solve problem 8, I would look up a proof that e is transcendental and repackage it so that it only seems to cover the special case. But that would be lame...
I also do not know how to solve #8 directly
If anybody could recover any theorem in five minutes then math would not be needed. There are classical facts which are easy to prove such as Gronwall inequality (Probl 3) and there are nontrivial facts. For example I can not imagine that anybody will solve #5 ex nihilo. But it easily follows from the standard ODE theorysorry for offtop
7. Prove the following well known theorem by using topological and analytical tools only. For every real symmetric matrix ##A## there is a real orthogonal matrix ##Q## such that ##Q^\tau AQ## is diagonal.
my read is compactness plus well chosen inequalities are fine. Is it fair game to use that if ##A=
\begin{bmatrix}
a_{1,1}& a_{1,2}\\
a_{1,2} & a_{2,2}\\
\end{bmatrix}
##
with positive diagonals, then ##a_{1,1}\cdot a_{2,2} -a_{1,2}^2\leq a_{1,1}\cdot a_{2,2} ## with equality iff ##A## is diagonal? I ask because this is a determinant inequality (Hadarmard's, though I only want the obvious 2x2 case) and perhaps 2x2 determinants are out of bounds.
my read is compactness plus well chosen inequalities are fine. Is it fair game to use that if ##A=
\begin{bmatrix}
a_{1,1}& a_{1,2}\\
a_{1,2} & a_{2,2}\\
\end{bmatrix}
##
with positive diagonals, then ##a_{1,1}\cdot a_{2,2} -a_{1,2}^2\leq a_{1,1}\cdot a_{2,2} ## with equality iff ##A## is diagonal? I ask because this is a determinant inequality (Hadarmard's, though I only want the obvious 2x2 case) and perhaps 2x2 determinants are out of bounds.
I don't see how you get along with ##n=2##, i.e. I fear I'll have to read a nasty induction ...
Nevertheless, coordinates will be necessary and the off diagonal elements are the key.
I don't see how you get along with ##n=2##, i.e. I fear I'll have to read a nasty induction ...
Nevertheless, coordinates will be necessary and the off diagonal elements are the key.
I am aware of the compactness proof of Wilf but I find all the indices involved to be unpleasant... so I am trying to figure out a slicker approach that yes hides a lot of ugliness via a (nice?) induction.
... so I am trying to figure out a slicker approach that yes hides a lot of ugliness via a (nice?) induction.
... c'mon, one lousy rotation ...
Maybe you can wait until the 7th before posting it. Would be a nice reminiscence!
#37
suremarc
147
65
fresh_42 said:
10. Let ##f(x)=2x^5-6x+6\in \mathbb{Z}[x]##. In which of the following rings is ##f## irreducible and why? (a) ##\mathbb{Z}[x]## (b) ##(S^{-1}\mathbb{Z})[x]## with ##S=\{2^n\,|\,n\in \mathbb{N}_0\}## (c) ##\mathbb{Q}[x]## (d) ##\mathbb{R}[x]## (e) ##\mathbb{C}[x]##
(abc). By Eisenstein's criterion, ##f(x)## is irreducible over ##\mathbb{Q}##. Proof: Choose p = 3; p divides each non-leading coefficient (-6, 6), but does not divide the leading coefficient (2). Furthermore, p^2 does not divide the constant term (6).
Since ##\mathbb{Z}## and ##S^{-1}\mathbb{Z}## are subrings of ##\mathbb{Q}##, it follows that ##f(x)## is irreducible over them as well.
(d). Compute the discriminant of ##f(x)##, for example via row reduction of the Sylvester matrix. The discriminant is then ##12^4\cdot (5^5 - 3\cdot 4^4) = 12^4 \cdot 2357##. Since the discriminant is positive, all roots are real, and ##f(x)## splits into linear factors over ##\mathbb{R}[x]##.
(e). By the fundamental theorem of algebra, every non-constant univariate polynomial with coefficients in ##\mathbb{C}## splits into linear factors over ##\mathbb{C}[x]##.
(abc). By Eisenstein's criterion, ##f(x)## is irreducible over ##\mathbb{Q}##. Proof: Choose p = 3; p divides each non-leading coefficient (-6, 6), but does not divide the leading coefficient (2). Furthermore, p^2 does not divide the constant term (6).
Since ##\mathbb{Z}## and ##S^{-1}\mathbb{Z}## are subrings of ##\mathbb{Q}##, it follows that ##f(x)## is irreducible over them as well.
(d). Compute the discriminant of ##f(x)##, for example via row reduction of the Sylvester matrix. The discriminant is then ##12^4\cdot (5^5 - 3\cdot 4^4) = 12^4 \cdot 2357##. Since the discriminant is positive, all roots are real, and ##f(x)## splits into linear factors over ##\mathbb{R}[x]##.
(e). By the fundamental theorem of algebra, every non-constant univariate polynomial with coefficients in ##\mathbb{C}## splits into linear factors over ##\mathbb{C}[x]##.
You have been sloppy, which made one answer wrong. (d) can be answered easier, if we only look at ##x\to \pm\infty##.
#39
suremarc
147
65
fresh_42 said:
You have been sloppy, which made one answer wrong. (d) can be answered easier, if we only look at ##x\to \pm\infty##.
Is it because 2 is irreducible in ##\mathbb{Z}##, and so is technically an irreducible 0-degree "polynomial"? Now I understand why the theorems I read mentioned the non-primitive case. Oh, and yeah, I guess since ##f## is an odd-degree polynomial, it has to have one real root. D'oh.
Is it because 2 is irreducible in ##\mathbb{Z}##, and so is technically an irreducible 0-degree "polynomial"? Now I understand why the theorems I read mentioned the non-primitive case. Oh, and yeah, I guess since ##f## is an odd-degree polynomial, it has to have one real root. D'oh.
Yes, ##2## isn't a unit in ##\mathbb{Z}## so ##2x^5-6x+6=2(x^5-3x+3)## is a legitimate split. Eisenstein makes is irreducible over ##\mathbb{Q}## and ##S## in part (b) makes it a unit, too, but not in part (a).
Problem 1 is a strange one. One readily finds a small dimensional counterexample to first bullet point. E.g -1 x = 1 is not solvable for x\geq 0. Am I missing something?
Problem 1 is a strange one. One readily finds a small dimensional counterexample to first bullet point. E.g -1 x = 1 is not solvable for x\geq 0. Am I missing something?
Maybe ...
fresh_42 said:
... exactly one of the following two statements is true ...
8. We define ##e=\displaystyle{\sum_{k=0}^\infty \dfrac{1}{k!}}.## Prove that ##e^2## is irrational.
Obviously the rationality of ##e^2## depends on that of ##e##: this has been done, just wanted to note that this proof was non-trivial in the sense that there exists an irrational number, say ##\sqrt{2}##, such that it's square (namely 2) is rational.
One readily verifies the subset O(n) \subseteq GL(n,\mathbb R) of orthogonal matrices is a group. A matrix is orthogonal iff its column vectors are pairwise orthonormal, therefore O(n) is bounded. The map f(A)=A^tA is continuous, therefore O(n) = f^{-1}(E) is closed (because singletons are closed). By Heine-Borel O(n) is compact.
Let A be a real n\times n matrix. Denote
<br />
\mathcal V(A) := \sum _{i\neq j} a_{ij}^2.<br />
I.e sums of squares not on the diagonal.
Iirc, this lemma is by Jacobi, but might be mistaken.
Statement. Let A be a real symmetric matrix. If \mathcal V(A) >0, then there exists a real orthogonal matrix B such that \mathcal V(B^tAB)<\mathcal V(A).
Sketch of Proof. Let a_{k\ell} \neq 0, where k\neq \ell. Choose B in the following manner. Start with E and put b_{kk} = b_{\ell\ell} = \cos\alpha and b_{k\ell} = -b_{\ell k} = \sin\alpha for some suitable \alpha. Clearly, B\in O(n).
Put C:= B^tAB. Then
<br />
c_{k\ell} = \sum _{i,j} b_{ik}a_{ij}b_{j\ell}<br />
For s,t\notin \{k,\ell\} we have c_{st} = a_{st}. Otherwise, for s\notin \{k,\ell\}
<br />
c_{sk} = a_{sk}\cos\alpha - a_{s\ell}\sin\alpha<br />
and
<br />
c_{s\ell} = a_{s\ell}\cos\alpha + a_{sk}\sin\alpha.<br />
Analogously for t\notin \{k,\ell\}. We then have
<br />
c_{sk}^2 + c_{s\ell}^2 = a_{sk}^2 + a_{e\ell}^2\ \ (s\notin \{k,\ell\}) \quad\mbox{and}\quad c_{kt}^2 + c_{\ell t}^2 = a_{kt}^2 + a_{\ell t}^2 \ \ (t\notin \{k,\ell\}).<br />
Now all that's left is to pick \alpha such that c_{k\ell} = 0, then \mathcal V(B^tAB) = \mathcal V(A) - 2a_{k\ell}^2 < \mathcal V(A). Note that
<br />
c_{k\ell} = (a_{kk} - a_{\ell\ell})\sin\alpha\cos\alpha + a_{k\ell}\cos 2\alpha.<br />
By IVT c_{k\ell} = 0 for some 0<\alpha<\pi/2.
Let A be a real symmetric matrix. Consider h:O(n) \to \mathrm{Mat}_n,\quad X\mapsto X^tAX. Since this map is continuous, the image \mathrm{im}(h) is compact. Thus, \mathcal V : \mathrm{im}(h) \to \mathbb R attains minimum on, say, M\in\mathrm{im}(h).
Put M=Q^tAQ for some Q\in O(n). Suppose \mathcal V(M)>0. Then for some orthogonal matrix B
<br />
\mathcal V((QB)^tAQB) < \mathcal V (Q^tAQ),<br />
which is impossible. So M must be a diagonal matrix and A is diagonalisable.
One readily verifies the subset O(n) \subseteq GL(n,\mathbb R) of orthogonal matrices is a group. A matrix is orthogonal iff its column vectors are pairwise orthonormal, therefore O(n) is bounded. The map f(A)=A^tA is continuous, therefore O(n) = f^{-1}(E) is closed (because singletons are closed). By Heine-Borel O(n) is compact.
Let A be a real n\times n matrix. Denote
<br />
\mathcal V(A) := \sum _{i\neq j} a_{ij}^2.<br />
I.e sums of squares not on the diagonal.
Iirc, this lemma is by Jacobi, but might be mistaken.
Statement. Let A be a real symmetric matrix. If \mathcal V(A) >0, then there exists a real orthogonal matrix B such that \mathcal V(B^tAB)<\mathcal V(A).
Sketch of Proof. Let a_{k\ell} \neq 0, where k\neq \ell. Choose B in the following manner. Start with E and put b_{kk} = b_{\ell\ell} = \cos\alpha and b_{k\ell} = -b_{\ell k} = \sin\alpha for some suitable \alpha. Clearly, B\in O(n).
Put C:= B^tAB. Then
<br />
c_{k\ell} = \sum _{i,j} b_{ik}a_{ij}b_{j\ell}<br />
For s,t\notin \{k,\ell\} we have c_{st} = a_{st}. Otherwise, for s\notin \{k,\ell\}
<br />
c_{sk} = a_{sk}\cos\alpha - a_{s\ell}\sin\alpha<br />
and
<br />
c_{s\ell} = a_{s\ell}\cos\alpha + a_{sk}\sin\alpha.<br />
Analogously for t\notin \{k,\ell\}. We then have
<br />
c_{sk}^2 + c_{s\ell}^2 = a_{sk}^2 + a_{e\ell}^2\ \ (s\notin \{k,\ell\}) \quad\mbox{and}\quad c_{kt}^2 + c_{\ell t}^2 = a_{kt}^2 + a_{\ell t}^2 \ \ (t\notin \{k,\ell\}).<br />
Now all that's left is to pick \alpha such that c_{k\ell} = 0, then \mathcal V(B^tAB) = \mathcal V(A) - 2a_{k\ell}^2 < \mathcal V(A). Note that
<br />
c_{k\ell} = (a_{kk} - a_{\ell\ell})\sin\alpha\cos\alpha + a_{k\ell}\cos 2\alpha.<br />
By IVT c_{k\ell} = 0 for some 0<\alpha<\pi/2.
Let A be a real symmetric matrix. Consider h:O(n) \to \mathrm{Mat}_n,\quad X\mapsto X^tAX. Since this map is continuous, the image \mathrm{im}(h) is compact. Thus, \mathcal V : \mathrm{im}(h) \to \mathbb R attains minimum on, say, M\in\mathrm{im}(h).
Put M=Q^tAQ for some Q\in O(n). Suppose \mathcal V(M)>0. Then for some orthogonal matrix B
<br />
\mathcal V((QB)^tAQB) < \mathcal V (Q^tAQ),<br />
which is impossible. So M must be a diagonal matrix and A is diagonalisable.
This proof is from Herb Wilf who passed away 9 years ago that day (*). I like how its pure analytic nature proves a statement from linear algebra. It shows that proof techniques do not necessarily belong to the area the statement stems from. Another famous example is the fundamental theorem of algebra which has various proofs.
Problem 4.
I solve the PDE$$
xu_x+yu_y+(x^2 +y^2)u_z=0
$$
by the "method of characteristics". Dividing by ##x## the PDE becomes
$$
u_x + \frac{y}{x}u_y + \frac{(x^2 + y^2)}{x}u_z=0
$$
The characteristic ODEs are,
$$
\frac{dx}{ds}=1
$$
$$
\frac{dy}{ds}=\frac{y}{x}
$$
$$
\frac{dz}{ds}=\frac{(x^2 + y^2)}{x}
$$
where ##s## is the characteristic parameter. Solving these ODEs we find,
$$
x=s
$$
$$
y=y_0x
$$
$$
z=z_0 + \frac{1}{2}(x^2 + y^2)
$$
The solution is any function of the constant curves,
$$
u(x,y,z)=F(y_0,z_0)
$$
Solving for ##y_0## and ##z_0## I find,
$$u(x,y,z)=F(\frac{y}{x},z-\frac{1}{2}(x^2 + y^2))
$$
In anticipation of the initial conditions I displace the ##z## coordinate by ##\frac{\pi}{2}## and guess the solution.
$$
u(x,y,z)=\frac{y}{x}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )
$$
Taking the partial derivatives I find,
$$
u_x=-\frac{y}{x^2}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )-y
$$
$$
u_y=\frac{1}{x}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )-\frac{y^2}{x}
$$
$$
u_z=\frac{y}{x}
$$
This function satisfies the initial conditions and the original PDE.
Problem 4.
I solve the PDE$$
xu_x+yu_y+(x^2 +y^2)u_z=0
$$
by the "method of characteristics". Dividing by ##x## the PDE becomes
$$
u_x + \frac{y}{x}u_y + \frac{(x^2 + y^2)}{x}u_z=0
$$
The characteristic ODEs are,
$$
\frac{dx}{ds}=1
$$
$$
\frac{dy}{ds}=\frac{y}{x}
$$
$$
\frac{dz}{ds}=\frac{(x^2 + y^2)}{x}
$$
where ##s## is the characteristic parameter. Solving these ODEs we find,
$$
x=s
$$
$$
y=y_0x
$$
$$
z=z_0 + \frac{1}{2}(x^2 + y^2)
$$
The solution is any function of the constant curves,
$$
u(x,y,z)=F(y_0,z_0)
$$
Solving for ##y_0## and ##z_0## I find,
$$u(x,y,z)=F(\frac{y}{x},z-\frac{1}{2}(x^2 + y^2))
$$
In anticipation of the initial conditions I displace the ##z## coordinate by ##\frac{\pi}{2}## and guess the solution.
$$
u(x,y,z)=\frac{y}{x}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )
$$
Taking the partial derivatives I find,
$$
u_x=-\frac{y}{x^2}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )-y
$$
$$
u_y=\frac{1}{x}\left ( z-\frac{\pi}{2}-\frac{1}{2}(x^2 + y^2)\right )-\frac{y^2}{x}
$$
$$
u_z=\frac{y}{x}
$$
This function satisfies the initial conditions and the original PDE.
Well, ##F## should be differentiable, not just any. We have infinitely many possible solutions and the initial values are useless. Those initial values given here corresponded to ##F(a,b)=a^2\sin(b),## but this is not a unique solution, as your solution shows.
My contribution to problem 7. Lots of simple inequalities and few indices, which is more my style. The little lemmas also give some other valuable information about symmetric matrices, via very simple means.
Let ##A## be a real symmetric matrix in all cases. And ##\mathbf e_k## denotes the relevant kth standard basis vector.
lemma 1:
a ##2\times 2## real symmetric ##A## with quadratic form
##f: \mathbb R_{\big \Vert \mathbb x\big \Vert_2=1}^2 \longrightarrow \mathbb R## given by ##f\big(\mathbf x\big)=\mathbf x^T A \mathbf x## (notice the domain is the unit circle), attains a maximum, ##m_x##
proof:
##f## is continuous, applied to a compact set, so its image is compact i.e. in this case closed and bounded in ##\mathbb R##, so it attains a maximum, ##m_x##.
lemma 2:
If ##A## is a ##2\times 2##, non-diagonal matrix with (at least) one zero on the diagonal, then ##m_x \gt 0##.
proof:
WLOG suppose ##a_{1,1}=0## (if not, re-label and run the argument on ##A':= P^TAP=PAP## where ##P## is the (non-identity) type 2 elementary matrix)
##f\big(\mathbf e_1\big) = 0##
now ##m_x \leq 0\implies f\big(\mathbf x_{m_x}\big) \leq 0 \implies \sigma^2 \cdot f\big( \mathbf x_{m_x}\big) =f\big(\sigma \cdot \mathbf x_{m_x}\big) \leq 0##
thus for notational ease we relax the length one constraint on ##\mathbf x## for this lemma only (i.e. let the domain be all of ##\mathbb R^2## for this lemma only)
so consider ##\mathbf x:=\mathbf e_1 + \beta \mathbf e_2##
##f\big(\mathbf x\big) = \mathbf x^TA\mathbf x = 2\cdot a_{1,2}\cdot x_{1}\cdot x_{2} + a_{2,2}\cdot x_{2}^2 =2\cdot \beta \cdot a_{1,2}+ \beta^2 a_{2,2} = \beta\big(\beta a_{2,2} + a_{1,2}\big) ##
i.) if ##a_{1,2} \gt 0## then the RHS is necessarily positive for small enough modulus ##\beta \gt 0##
(i.e. if ##a_{2,2}=0## then any positive ##\beta## works, otherwise we have
##\beta a_{2,2} + a_{1,2}\geq a_{1,2}-\beta \big \vert a_{2,2}\big \vert\geq \frac{1}{2}a_{1,2}\gt 0##
by selecting ##\beta \in \big(0, \frac{a_{1,2}}{2 \vert a_{2,2}\vert }\big)##
ii.) similarly if ##a_{1,2} \lt 0## then the RHS is necessarily positive for small enough modulus ##\beta \lt 0##
iii.) ##a_{1,2} \neq 0## because ##A## is symmetric but not diagonal
this complete the proof of the lemma, i.e. ##m_x \gt 0##
For ##2\times 2## real symmetric ##A##, ##m_x \geq \max\big(a_{1,1},a_{2,2}\big)## with equality *iff* ##A## is diagonal.
For the first direction: if ##A## is diagonal, then
##m_x=f\big(\mathbf x_{m_x}\big) = a_{1,1}\cdot x_{1}^2 + a_{2,2}\cdot x_{2}^2\leq \max\big(a_{1,1},a_{2,2}\big)\cdot \big(x_1^2 + x_2^2)= \max\big(a_{1,1},a_{2,2}\big)##
and ##\max\Big(f\big(\mathbf e_{1}\big), f\big(\mathbf e_{2}\big)\Big) =\max\big(a_{1,1},a_{2,2}\big)\leq m_x ## so this is met with equality.
For the other direction, suppose ##A## is not diagonal. Then ##B:=A-\max\big(a_{1,1},a_{2,2}\big)\cdot I##
is symmetric ##2\times 2## but not diagonal, with a zero on the diagonal. Applying the prior lemma to ##B## tells us that the maximal quadratic form is positive. i.e. there must exist (length one) ##\mathbf x## where
##0 \lt \mathbf x^T \Big(A-\max\big(a_{1,1},a_{2,2}\big)\cdot I\Big)\mathbf x = f\big(\mathbf x\big) - \max\big(a_{1,1},a_{2,2}\big)\mathbf x^T \mathbf x = f\big(\mathbf x\big)-\max\big(a_{1,1},a_{2,2}\big) \leq m_x -\max\big(a_{1,1},a_{2,2}\big)##
##\implies \max\big(a_{1,1},a_{2,2}\big) \lt m_x##
(where ##m_x## refers to maximum value of the quadratic form with respect to A, not B, i.e ##\mathbf x^T A \mathbf x##)
**now assume** ##A## is ##n\times n##
##f\big(\mathbf x\big)=\mathbf x^T A \mathbf x## (again with the constraint of ##\Big \Vert \mathbf x\big \Vert_2 = 1##) attains a maximum ##m_x## again for reasons of applying a continuous function to a compact domain, with the real line as the codomain. Define orthogonal ##Q_1## to have its first col vector ##\mathbf x_{m_x}## (i.e. some unit length vector such that ##f\big(\mathbf x_{m_x}\big) = m_x## and extend to an orthonormal basis.)
##B:= Q_1^T A Q_1 = \begin{bmatrix}
m_x & *^T\\
* & A_{n-1\times n-1}\\
\end{bmatrix}##
note: ##B## is symmetric, so is ##A_{n-1\times n-1}## and we claim ##*=\mathbf 0_{n-1}##. Suppose this is not the case and there is ##b_{i,1}=\gamma \neq 0## for some ##i\in \big\{2,3,...,n\big\}##. Let ##P## be the elementary type 2 matrix that swaps element i with element 2 (which is a symmetric permutation matrix).
Then ##P^TBP## has a leading ##2\times 2## principal submatrix given by
##\begin{bmatrix}
m_x & \gamma\\
\gamma & b_{i,i}\\ \end{bmatrix} ##
By the lemma 3, since this ##2 \times 2## symmetric matrix is not diagonal there is some maximizing ##\mathbf x## on the unit circle such that ##\eta =\mathbf x^T\begin{bmatrix}
m_x & \gamma\\
\gamma & *\\\end{bmatrix} \mathbf x \gt \max\big(b_{1,1},b_{i,i}\big)= \max\big(m_x,*\big)\geq m_x##
Finally, define
##\mathbf v :=Q_1P\begin{bmatrix}
\mathbf x \\
\mathbf 0_{n-2}\\\end{bmatrix} ##, noting that ##\Big \Vert \mathbf v\Big \Vert_2 = 1## because ##Q_1## and ##P## are orthogonal
thus ##\mathbf v^T A\mathbf v = \eta \gt m_x## which is a contradiction.
Thus we've proven
##B= Q_1^T A Q_1 = \begin{bmatrix}
m_x & \mathbf 0_{n-1}^T\\
\mathbf 0_{n-1} & A_{n-1\times n-1}\\
\end{bmatrix}## and by inductive hypothesis (with the trivial base case of ##A_{n-1\times n-1}## being ##1\times 1##) there is an orthogonal matrix ##Q_{n-1\times n-1}## such that ##Q_{n-1\times n-1}^T A_{n-1\times n-1} Q_{n-1\times n-1} = D_{n-1}## so
##Q_2:= \begin{bmatrix}
1 & \mathbf 0_{n-1}^T\\
\mathbf 0_{n-1} & Q_{n-1\times n-1}\\
\end{bmatrix}##
##Q:= Q_1Q_2## and ##Q^TAQ = \begin{bmatrix}
m_x & \mathbf 0_{n-1}^T\\
\mathbf 0_{n-1} & D_{n-1\times n-1}\\
\end{bmatrix}=D## as desired
lemma 2:
If ##A## is a ##2\times 2##, non-diagonal matrix with (at least) one zero on the diagonal, then ##m_x \gt 0##.
proof:
WLOG suppose ##a_{1,1}=0## (if not, re-label and run the argument on ##A':= P^TAP=PAP## where ##P## is the (non-identity) type 2 elementary matrix)
##f\big(\mathbf e_1\big) = 0##
Why isn't the proof done here? If we have a point with ##f(\mathbf{e_1})=0,## isn't the maximum ##f(\mathbf{x}_{m_x})## automatically non-negative? The case ##f(\mathbf{x}_{m_x})=0## is easy to exclude.
Why isn't the proof done here? If we have a point with ##f(\mathbf{e_1})=0,## isn't the maximum ##f(\mathbf{x}_{m_x})## automatically non-negative? The case ##f(\mathbf{x}_{m_x})=0## is easy to exclude.
Agreed that non-negativity is immediate. I wanted to exclude ##f(\mathbf{x}_{m_x})=0## in a methodical way (without using eigenvectors, bilinear forms, etc.). How would you do it?
Agreed that non-negativity is immediate. I wanted to exclude ##f(\mathbf{x}_{m_x})=0## in a methodical way (without using eigenvectors, bilinear forms, etc.). How would you do it?
If ##a_{22}\neq 0## then ##\mathbf{x}=(0,1)## is positive, and if ##a_{22}=0 \wedge a_{12}\neq 0## then ##\mathbf{x}=(a_{12}^{-1},1)## does it, and if ##a_{22}=0 \wedge a_{12}= 0## then ##A=0## in which case there is nothing to do. But you do not need to make the cases, it can already be seen from the equation for ##f(\mathbf{x})##, even without extending the domain.
Would Liouville work? O.k. it is shooting sparrows with canons, but ...
If ##a_{22}\neq 0## then ##\mathbf{x}=(0,1)## is positive...
I was thinking of ##a_{2,2} \lt 0## so ##f(\mathbf e_2)\lt 0##...
the fact that
##\beta\big(\beta a_{2,2} + a_{1,2}\big) ##
equals 0 with ##\beta =0## and the sign changes in some small ball around zero had nice analytic feel to it since it basically becomes ##\beta\cdot a_{1,2}## in that neighborhood.
fresh_42 said:
Would Liouville work? O.k. it is shooting sparrows with canons, but ...
I originally had a mind to lift something from complex analysis since ##\mathbb R^2## is a bit like ##\mathbb C## though I didn't see holomorphicity. There maybe be some cannons that can be used though,
I'm tired and want to turn in. Here's a go at Problem 3:
We are given:
\begin{align*}
z(t) \leq C + L \left| \int_{t_0}^t z (t_1) dt_1 \right| .
\end{align*}
Note that because ##z(t)## is a non-negative function on ##[a,b]##, and since the above inequality is required to hold for all ##t \in [a,b]##, this is why we it is required that ##C \geq 0##.
First assume that ##t \geq t_0##, so we can write:
\begin{align*}
z(t) \leq C + L \int_{t_0}^t z (t_1) dt_1 .
\end{align*}Since ##L \geq 0## we may substitute the inequality into the RHS which gives a first iteration:
\begin{align*}
z(t) & \leq C+ L \int_{t_0}^t \left( C + L \int_{t_0}^{t_1} z (t_2) dt_2 \right) dt_1 \\
& = C + C L \int_{t_0}^t + L^2 \int_{t_0}^t \int_{t_0}^{t_1} z (t_2) dt_2 dt_1
\end{align*}
A second iteration gives:
\begin{align*}
z(t) & \leq C + C L \int_{t_0}^t + L^2 \int_{t_0}^t \int_{t_0}^{t_1} \left( C + L \int_{t_0}^{t_2} z (t_3) dt_3 \right) dt_2 dt_1 \\
& = C + C L \int_{t_0}^t + C L^2 \int_{t_0}^t \int_{t_0}^{t_1} dt_2 dt_1 + L^3 \int_{t_0}^t \int_{t_0}^{t_1} \int_{t_0}^{t_2} z (t_3) dt_3 dt_2 dt_1
\end{align*}
An ##n##th iteration gives
\begin{align*}
z(t) & \leq C + C \sum_{i=1}^n A^{(i)} + L^{n+1} \int_{t_0}^t \int_{t_0}^{t_1} \int_{t_0}^{t_2} \cdots \int_{t_0}^{t_n} z (t_{n+1}) dt_{n+1} dt_n dt_{n-1} \dots dt_1 \quad (*)
\end{align*}
\begin{align*}
t \leq t_n \leq t_{n-1} \leq \cdots \leq t_1 \leq t .
\end{align*}As ##z(t)## is non-negative real valued continuous function over the closed interval ##[a,b]## it attains its maximum value, denoted by ##z_\max##. We deal with the last term in (*):
I'm tired and want to turn in. Here's a go at Problem 3:
We are given:
\begin{align*}
z(t) \leq C + L \left| \int_{t_0}^t z (t_1) dt_1 \right| .
\end{align*}
Note that because ##z(t)## is a non-negative function on ##[a,b]##, and since the above inequality is required to hold for all ##t \in [a,b]##, this is why we it is required that ##C \geq 0##.
First assume that ##t \geq t_0##, so we can write:
\begin{align*}
z(t) \leq C + L \int_{t_0}^t z (t_1) dt_1 .
\end{align*}Since ##L \geq 0## we may substitute the inequality into the RHS which gives a first iteration:
\begin{align*}
z(t) & \leq C+ L \int_{t_0}^t \left( C + L \int_{t_0}^{t_1} z (t_2) dt_2 \right) dt_1 \\
& = C + C L \int_{t_0}^t + L^2 \int_{t_0}^t \int_{t_0}^{t_1} z (t_2) dt_2 dt_1
\end{align*}
A second iteration gives:
\begin{align*}
z(t) & \leq C + C L \int_{t_0}^t + L^2 \int_{t_0}^t \int_{t_0}^{t_1} \left( C + L \int_{t_0}^{t_2} z (t_3) dt_3 \right) dt_2 dt_1 \\
& = C + C L \int_{t_0}^t + C L^2 \int_{t_0}^t \int_{t_0}^{t_1} dt_2 dt_1 + L^3 \int_{t_0}^t \int_{t_0}^{t_1} \int_{t_0}^{t_2} z (t_3) dt_3 dt_2 dt_1
\end{align*}
An ##n##th iteration gives
\begin{align*}
z(t) & \leq C + C \sum_{i=1}^n A^{(i)} + L^{n+1} \int_{t_0}^t \int_{t_0}^{t_1} \int_{t_0}^{t_2} \cdots \int_{t_0}^{t_n} z (t_{n+1}) dt_{n+1} dt_n dt_{n-1} \dots dt_1 \quad (*)
\end{align*}
\begin{align*}
t \leq t_n \leq t_{n-1} \leq \cdots \leq t_1 \leq t .
\end{align*}As ##z(t)## is real valued continuous function over the closed interval ##[a,b]## it attains its maximum value, denoted by ##z_\max##. We deal with the last term in (*):
A remark that this follows from the convergence of the Taylor series of the exponential function would had been sufficient. Instead you should have proven the iterated integral which you used twice, or at least mentioned that it is the volume formula for parallelepipeds.
Finally we have
\begin{align*}
z(t) & \leq C + C \sum_{i=1}^\infty A^{(i)} \\
& = C + C \sum_{i=1}^\infty \frac{1}{i!} L^i (t - t_0)^i \\
& = C e^{L (t - t_0)}
\end{align*}Now assume that ##t_0 > t##, then
\begin{align*}
z(t) \leq C+ L \left| \int_{t_0}^t z (t_1) dt_1 \right| = C + L \int_t^{t_0} z (t_1) dt_1
\end{align*}
and giving the same argument again we obtain
\begin{align*}
z(t) \leq C e^{L (t_0 - t)} = C e^{L |t - t_0|} .
\end{align*}
Well done. You could have avoided the entire iteration, if you had used a function ##F## and considered ##\dfrac{F'}{F}\leq L##. This is as good as the exponential function and shifts the iteration into the derivative.
Imagine a thread with this topic placed in our Topology and Analysis forum rather than in the Linear And Abstract Algebra forum where it perfectly fits without being moved. The first sentences could be:
The orthogonal real matrices build a compact subset of the Euclidean real space, because ... Now we define a continuous function ...
The present formulation of Problem 5 makes it look like the path ##x## is also prescribed, while I believe you want the solver to prove that such a path ##x## exists and is unique for a prescribed ##A## and a prescribed initial condition. This (and the typo that was already corrected) was what I hinted at.
The present formulation of Problem 5 makes it look like the path ##x## is also prescribed, while I believe you want the solver to prove that such a path ##x## exists and is unique for a prescribed ##A## and a prescribed initial condition. This (and the typo that was already corrected) was what I hinted at.
The problem means: Solve the differential equation system ##\dot x=Ax## and show that ##x(0)=x_0## makes the solution unique. (I added this to the problem, thanks.)