 Download Presentation Chapter 5 Solving Linear Equation Systems

# Chapter 5 Solving Linear Equation Systems

Download Presentation ## Chapter 5 Solving Linear Equation Systems

- - - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - - -
##### Presentation Transcript

1. Chapter 5 Solving Linear Equation Systems

2. Objectives • Understand the importance of solving linear equation • Know the Gauss Elimination method • Know LU decomposition for determine matrix inversion • Know the Gauss Seidel method

3. Content • Introduction • Gauss Elimination • LU Decomposition and Matrix Inversion • Gauss Seidel • Conclusions

4. Introduction (1) Linear equation in n variables a11x1+a12x2+ … +a1nxn = b1 a21x1+a22x2+ … +a2nxn = b2 a31x1+a32x2+ … +a3nxn = b3 an1x1+an2x2+ … +annxn = bn Must solve with computer !!

5. Introduction (2) Solving Small Numbers of Equations • There are many ways to solve a system of linear equations: • Graphical method • Cramer’s rule • Method of elimination • Computer methods For n ≤ 3

6. Introduction (3) Graphical method slope intercept

7. Introduction (4) Graphical method pitfalls

8. Introduction (5) Cramer’s rule Ratios of determinants of the array of coefficients of the equations.

9. Gauss Elimination (1) Principle: Successively solve one of the equations of the set for one of the unknowns and eliminate that variable from the remaining equations by substitution. Applying to computer method: Developing a systematic scheme or algorithm to eliminate unknowns and to back substitute. Two main algorithm steps : Forward elimination and back substitution

10. Gauss Elimination (2)

11. Gauss Elimination (3) • Division by zero. • Round-off errors. • Ill-conditioned systems (two or more equations are nearly identical) • Singular systems (at least two equations are identical) Pitfalls

12. Gauss Elimination (4) Gauss Jordan (Variation of Gauss Elimination) • Elimination from other equations rather than subsequent ones • All rows are normalised by dividing them by their pivot elements • Eliminating step result in an identity matrix • No necessary for back substitution process

13. LU Decompos.. (1) • An efficient way to compute matrix inverse L and U are the Lower and Upper triangular matrices

14. LU Decompos.. (2) • Pros LU decomposition • requires the same total FLOPS as for Gauss elimination. • Saves computing time by separating time-consuming elimination step from the manipulations of the right hand side. • Provides efficient means to compute the matrix inverse

15. LU Decompos.. (3) • System condition Inverse of a matrix provides a means to test whether systems are ill-conditioned. Check matrix condition by The norm of matrix can be computed from

16. LU Decompos.. (4) • Relative error • The relative error of the norm of the computed solution can be as large as the relative error of the norm of the coefficients of [A] multiplied by the condition number. • For example, if the coefficients of [A] are known to t-digit precision (rounding errors~10-t) and Cond [A]=10c, the solution [X] may be valid to only t-c digits (rounding errors~10c-t).

17. Gauss-Seidel(1) • Iterative or approximate methods provide an alternative to the elimination methods. The Gauss-Seidel method is the most commonly used iterative method. • The system [A]{X}={B} is reshaped by solving the first equation for x1, the second equation for x2, and the third for x3, …and nth equation for xn. For conciseness, we will limit ourselves to a 3x3 set of equations.

18. Gauss-Seidel(2) • Now we can start the solution process by choosing guesses for the x’s. A simple way to obtain initial guesses is to assume that they are zero. These zeros can be substituted into x1equation to calculate a new x1=b1/a11.

19. Gauss-Seidel(3) • New x1 is substituted to calculate x2 and x3. The procedure is repeated until the convergence criterion is satisfied: For all i, where j and j-1 are the present and previous iterations.

20. Gauss-Seidel(4)