Can Different Access Patterns Improve a Matrix's Condition Number?

In summary, the conversation discusses the possibility of improving the condition number of a symmetric positive-definite matrix by changing the access pattern and the effect of different access orders on the Eigenvalue Spectral Radius. It also explores the discrepancy in the number of iterations needed for convergence when executing an iterative algorithm on a GPU versus a CPU. The cause of this discrepancy is attributed to round-off error on the GPU and is typically solved by using double precision or the Kahan approximation.
  • #1
nurfherder
4
0
Hello all,

I am new to this forum but am glad I found it, I have a quick question about condition numbers and order of operations.

Given a symmetric positive-definitive matrix with a initial condition number α, is it possible to improve that condition number with a different access pattern? For example, if I access the matrix within the context of an iterative solver (e.g., Jacobin) in column-major order would it improve the condition number over access done in row-major order?

I am doing some personal research into iterative solvers and convergence rates and I would like to know if the condition number can be improved, thus lower the total number of iterations to converge, significantly with something so small.

Thank you.
 
Physics news on Phys.org
  • #2
Never mind about my initial post, the order of operations as applied to column major versus row major will have no effect on the condition number of the matrix. The access order should not effect the Eigenvalue Spectral Radius.

Does anyone have any clue as to why a iterative algorithm, such as Jacobin, would have less iterations to converge when executed on the GPU versus the CPU? The model and tolerance is exactly the same in both cases, so I cannot understand how the GPU has less iterations using a Krylov search space. I have executed the SAME code for CPU and GPU (except of course the CPU has NO calls to the GPU) on two different sets of CPUs and GPUs (one double precision - Tesla, and one not - Quadro) and get exactly the same result.

Any ideas would be great, I think I might have broke one of my neurons on this one.

Thanks.
 
  • #3
nurfherder said:
Does anyone have any clue as to why a iterative algorithm, such as Jacobin, would have less iterations to converge when executed on the GPU versus the CPU? The model and tolerance is exactly the same in both cases, so I cannot understand how the GPU has less iterations using a Krylov search space.

That doesn't make any sense to me. If you do the EXACT same operations, you should get exactly the same results.

The explanation may be something to do with compiler optimisation, compiler bugs, library routines, different implementations of floating point arithmetic, etc. The only way to nail that is compare your two calculations step by step. If there are differences, they will probably show up (if only in the last decimal place) on small matrices as well as on big ones.
 
  • #4
You are right - it doesn't make sense to me either. I was just wondering if there was an obvious and therefore easy reason.

Thank you for your help and time.
 
  • #5
I found the problem.

Turns out that the GPU typically has some round-off error that will benefit the GPU for iterative solvers such that the higher precision of the CPU will take more iterations to converge. The small inaccuracies of the GPU become magnified when doing large sets of summations - such as those found in the Dot-product of the iterative solver I am using.

It is sneaky and is typically solved by using double precision (CUDA arch. 1.3 or greater) or algorithmically with the Kahan approximation.
 
  • #6
nurfherder said:
Turns out that the GPU typically has some round-off error that will benefit the GPU for iterative solvers such that the higher precision of the CPU will take more iterations to converge.

Hm... long before the days of GPUs, I remember a CFD software guru trying to convince me that his code worked better in 32 bit precision arithmetic than in 64 (In fact it didn't work at all in 64, on most problems).

Maybe he gave up trying to sell his CFD software and went into GPU design ... :smile:
 

What is the Matrix Condition Number?

The Matrix Condition Number is a measure of the sensitivity of a matrix's output to changes in its input. It indicates how much the output of a matrix will change when there is a small change in its input.

How is the Matrix Condition Number calculated?

The Matrix Condition Number is calculated by taking the ratio of the largest singular value of the matrix to the smallest singular value. This value is a non-negative number that can range from 1 to infinity.

Why is the Matrix Condition Number important?

The Matrix Condition Number is important because it can indicate how stable or accurate a solution to a system of linear equations will be. A higher condition number means that small changes in the input can result in large changes in the output, making the solution less stable.

What does a high Matrix Condition Number indicate?

A high Matrix Condition Number indicates that the matrix is ill-conditioned, meaning it is sensitive to changes in its input and may produce less accurate or stable solutions. This can be problematic in numerical computations and can lead to errors.

How can the Matrix Condition Number be improved?

The Matrix Condition Number can be improved by using techniques such as regularization or re-scaling the matrix's input data. These methods can help reduce the sensitivity of the matrix and improve the accuracy and stability of its solutions.

Similar threads

  • Linear and Abstract Algebra
Replies
10
Views
2K
  • Linear and Abstract Algebra
Replies
2
Views
1K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
2
Views
1K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
2
Views
1K
  • Programming and Computer Science
Replies
1
Views
2K
  • General Math
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
11
Views
1K
  • Linear and Abstract Algebra
Replies
7
Views
2K
  • Atomic and Condensed Matter
Replies
4
Views
1K
  • Linear and Abstract Algebra
Replies
1
Views
6K
Back
Top