CSE 291 Numerical Methods Matrix Computations Iterative Methods

































- Slides: 33

CSE 291: Numerical Methods Matrix Computations: Iterative Methods I Chung-Kuan Cheng

Outline o Introduction o Direct Methods o Iterative Methods n n n Formulations Projection Methods Krylov Space Methods Preconditioned Iterations Multigrid Methods Domain Decomposition Methods 2

Introduction Iterative Methods Direct Method LU Decomposition Domain Decomposition General and Robust but can be complicated if N>= 1 M Preconditioning Conjugate Gradient Jacobi GMRES Multigrid Gauss-Seidel Excellent choice for SPD matrices Remain an art for arbitrary matrices 3

Introduction: Matrix Condition 4

Introduction: Matrix norm 5

Introduction: Scaling 6

Introduction: Gershgorin Circle Theorem 7

Gershgorin Circle Theorem: Example 2 0. 1 0. 2 0. 3 0. 1 3 0 0. 3 4 0. 1 0 0 0. 2 5 8

Iterative Methods o 9

Stationary: Jacobi Method 10

Stationary-Gause-Seidel 11

Stationary: Successive Overrelaxation (SOR) Devised by extrapolation applied to Gauss-Seidel in the form of weighted average: In matrix terms the method becomes: where D, -L and -U represent the diagonal, the strictly lowertrg and strictly upper-trg parts of M M=D-L-U 12

SOR o Choose w to accelerate the convergence n W =1 : Jacobi / Gauss-Seidel n 2>W>1: Over-Relaxation n W < 1: Under-Relaxation 13

Convergence of Stationary Method o 14

Convergence of Gauss-Seidel o 15

Linear Equation: an optimization problem o 16

Linear Equation: an optimization problem o 17

18

Gradient of quadratic form 19

o since We have, If p != x 20

If A is not positive definite a) Positive definite matrix b) negative-definite matrix c) Singular matrix d) positive indefinite matrix 21

Non-stationary Iterative Method o State from initial guess x 0, adjust it until close enough to the exact solution i=0, 1, 2, 3, …… Adjustment Direction Step Size o How to choose direction and step size? 22

Steepest Descent Method (1) o Choose the direction in which f decrease most quickly: the direction opposite of o Which is also the direction of residue 23

Steepest Descent Method (2) o How to choose step size ? n Line Search should minimize f, along the direction of , which means Orthogonal 24

Steepest Descent Algorithm Given x 0, iterate until residue is smaller than error tolerance 25

Steepest Descent Method: example a) Starting at (-2, -2) take the direction of steepest descent of f b) Find the point on the intersection of these two surfaces that minimize f c) Intersection of surfaces. d) The gradient at the bottommost point is orthogonal to the gradient of the previous step 26

Iterations of Steepest Descent Method 27

Convergence of Steepest Descent-1 Eigenvector: Eigen. Value: j=1, 2, …, n Energy norm: 28

Convergence of Steepest Descent-2 29

Convergence Study (n=2) assume let Spectral condition number let 30

31

Case Study 32

Bound of Convergence It can be proved that it is also valid for n>2, where 33