Gauss-Seidl, Gauss-Southwell, Jacobi

  • Context: Graduate 
  • Thread starter Thread starter onako
  • Start date Start date
  • Tags Tags
    Jacobi
Click For Summary

Discussion Overview

The discussion centers around numerical optimization techniques, specifically the Gauss-Seidl, Gauss-Southwell, and Jacobi methods. Participants explore the relationships between these methods, their applications in solving linear systems, and conditions for convergence, particularly in the context of quadratic functions.

Discussion Character

  • Exploratory
  • Technical explanation
  • Mathematical reasoning
  • Debate/contested

Main Points Raised

  • One participant questions whether the Gauss-Seidl method is synonymous with the cyclic descent method (cydm) and seeks clarification on their relationship.
  • Another participant presents a function and an update formula, asking how to demonstrate that the value of the function decreases after a Jacobi iteration under certain conditions.
  • A subsequent reply challenges the validity of the claim that the function value decreases, providing a counterexample with specific matrix and vector values.
  • Another participant points out that the matrix in the counterexample is not strictly diagonally dominant, which is a necessary condition for the convergence of the Jacobi method.
  • One participant suggests using the Gershgorin circle theorem to prove conditions related to the eigenvalues of the matrix involved in the Jacobi method.
  • A later reply discusses the structure of the matrix derived from the update formula and its implications for the eigenvalues, seeking further guidance on the proof process.
  • Another participant proposes a method to simplify the problem by shifting the origin to eliminate the vector b, suggesting a focus on the eigenvectors of the matrix in question.

Areas of Agreement / Disagreement

Participants express differing views on the relationship between the Gauss-Seidl method and cydm, as well as the conditions under which the Jacobi method guarantees a decrease in function value. The discussion remains unresolved regarding the implications of the eigenvalue conditions and the proof of function behavior after iterations.

Contextual Notes

Participants note the importance of strict diagonal dominance for the convergence of the Jacobi method, and there are references to specific mathematical properties and theorems that may influence the discussion's outcomes.

onako
Messages
86
Reaction score
0
Couple of days ago I downloaded a book on numerical optimization, hoping to get clearer picture on some techniques.
But, I'm surprised that some of the concepts were not precisely separated from one another.

Namely, in the part of "coordinate descent methods" (cdm), I found that, in the case the contribution of each point
to the function in question is known, one applied the Gauss-Southwell approach to select the point to be updated.
In case the update is done is a sequence from 1 to n, the the approach is referred to as the "cyclic descent method" (cydm).
However, on some other places I've found that Gauss-Seidl method's principle is close to cydm's.
Is Gauss-Seidl actually another name for cydm? I hope someone will clarify this for me.

Another question is regarding the Jacobi method, which is a "hold on" version of Gauss-Seidl, in my understanding
(regarding the point update at the end of iteration). Given certain function f(x) which is quadratic in x,
in order to obtain its minimal, I set the derivative to 0, and obtain the following linear system

Ax=b

which I decide to solve by Jacobi iteration. Given that A is strictly diagonally dominant, is it true that the value of
f(x1), where x1 is obtained by a single Jacobi iteration on some arbitrary x0 (but not the stationary point) satisfies
f(x1)<f(x0)?
 
Physics news on Phys.org
Let me clarify what I'm looking for : given a function
<br /> f(x)=\frac{1}{2}x^TAx - b^Tx,<br />
and the update
<br /> x_1=D^{-1}(b-Rx_0)<br />
where A=D+R, with the being the diagonal matrix with diagonal entries of A (and R containing the off-diagonal entries of A), how could I show
<br /> f(x_0)&gt;f(x_1)<br />
 
What you want to show is false, in general. For example take
##A = \begin{bmatrix}1 & 2 \\ 2 & 1 \end{bmatrix}, \quad
b = \begin{bmatrix} 0 \\ 0 \end{bmatrix}, \quad
x_0 = \begin{bmatrix} 1 \\ 1 \end{bmatrix}##
 
Your A is not strictly diagonally dominant (which is necessary for convergence of Jacobi method).
 
Many thanks for the suggestion. Note that following:
<br /> D^{-1}R=D^{-1}(V-D)=D^{-1}V-I<br />
In my choice of V, the matrix D^{-1}V
reduces to a symmetric matrix with the only non-negative elements on the diagonal begin equal to 1:
<br /> D^{-1}V=\left(\begin{array}{ccccc}<br /> 1 &amp; &amp; &amp; &amp; \\<br /> &amp; \ddots &amp; &amp; -c_{ij} \\<br /> &amp; &amp; \ddots &amp; &amp; \\<br /> &amp; -c_{ij} &amp; &amp; \ddots &amp; \\<br /> &amp; &amp; &amp; &amp; 1<br /> \end{array}\right) \in\mathbb{R}^{n\times n}<br />
such that the sum of the absolute values of all the off-diagonal elements in each row and column sum to 1, which is
the corresponding diagonal element. By matrix difference $D^{-1}V-I$, this implies that the diagonal elements of the matrix $D^{-1}R$ are zero, and in terms of the proposed Gershgorin circle theorem, this would correspond to
n disks D(0, 1), meaning that eigenvalues of eigs(D^{-1}R)\leq 1 (correct me if I'm wrong).
How to proceed with the proof further?
 
You can get rid of the vector b by shifting the origin so the solution vector is zero, i.e ##y = x - \bar x## where ##A \bar x - b = 0##.

Then use the fact that any vector y is a linear combination of the eigenvectors of ##D^{-1}R##.
 

Similar threads

  • · Replies 5 ·
Replies
5
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 1 ·
Replies
1
Views
4K
  • · Replies 5 ·
Replies
5
Views
7K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 8 ·
Replies
8
Views
2K