Linear Equations - PowerPoint PPT Presentation

1 / 64
About This Presentation
Title:

Linear Equations

Description:

Gauss-Seidel Method Improved method Comparison of different methods Gauss is slow convergent; however, more stable. Gauss-Seidel might not be stable; ... – PowerPoint PPT presentation

Number of Views:50
Avg rating:3.0/5.0
Slides: 65
Provided by: Stan115
Category:

less

Transcript and Presenter's Notes

Title: Linear Equations


1
Linear Equations
  • PSCI 702
  • September 21, 2005

2
Numerical Errors
  • Round off Error
  • Truncation Error

3
Linear Algebraic Equations
4
Linear Algebraic Equations
  • In Matrix Format
  • Solution

5
Cramers Rule
  • Mij is the determinant of the Matrix A with the
    ith row and jth column removed.
  • (-1)ij is called the cofactor of element aij.

6
Cramers Rule
7
Cramers Rule
8
Cramers Rule
  • 3n2 operations for the determinant.
  • 3n3 operations for every unknown.
  • Unstable for large Matrices.
  • Large error propagation.
  • Good for small Matrices (nlt20).

9
Gaussian Elimination
  • Divide each row by the leading element.
  • Subtract row 1 from all other rows.
  • Move to the second row an continue the process.

10
Gaussian Elimination
11
Gaussian Elimination
12
Gaussian Elimination
  • Back substitution

13
Gaussian Elimination
  • Division by zero May occur in the forward
    elimination steps.
  • Round-off error Prone to round-off errors.

14
Gaussian Elimination
Consider the system of equations Use five
significant figures with chopping


At the end of Forward Elimination


15
Gaussian Elimination
Back Substitution
16
Gaussian Elimination
Compare the calculated values with the exact
solution
17
Improvements
Increase the number of significant
digits Decreases round off error Does not avoid
division by zero
Gaussian Elimination with Partial
Pivoting Avoids division by zero Reduces round
off error
18
Partial Pivoting
Gaussian Elimination with partial pivoting
applies row switching to normal Gaussian
Elimination. How? At the beginning of the kth
step of forward elimination, find the maximum of
If the maximum of the values is
In the pth row,
then switch rows p and k.
19
Partial Pivoting
What does it Mean? Gaussian Elimination with
Partial Pivoting ensures that each step of
Forward Elimination is performed with the
pivoting element akk having the largest
absolute value.
20
Partial Pivoting Example
Consider the system of equations

In matrix form

Solve using Gaussian Elimination with Partial
Pivoting using five significant digits with
chopping
21
Partial Pivoting Example
Forward Elimination Step 1 Examining the values
of the first column 10, -3, and 5 or 10,
3, and 5 The largest absolute value is 10, which
means, to follow the rules of Partial Pivoting,
we switch row1 with row1.
Performing Forward Elimination
22
Partial Pivoting Example
Forward Elimination Step 2 Examining the values
of the first column -0.001 and 2.5 or 0.0001
and 2.5 The largest absolute value is 2.5, so row
2 is switched with row 3
Performing the row swap
23
Partial Pivoting Example
Forward Elimination Step 2 Performing the
Forward Elimination results in
24
Partial Pivoting Example
Back Substitution Solving the equations through
back substitution
25
Partial Pivoting Example
Compare the calculated and exact solution The
fact that they are equal is coincidence, but it
does illustrate the advantage of Partial Pivoting
26
Gauss Jordan Elimination
  • Start with the following system of Matrices.
  • Divide each row by the leading element.
  • Subtract row 1 from all other rows.
  • In addition to subtracting the line whose
    diagonal term has been made unity from all those
    bellow it, also subtract from the equations above
    it as well.

27
Gauss Jordan Elimination
28
Matrix Factorization
  • Assume A can be written as AVU where V and U are
    triangular Matrices.

29
Matrix Factorization
Proof
If solving a set of linear equations If
Then Multiply by Which gives Remember
which leads to Now, if
then Now, let Which ends
with and
(1) (2)
30
Matrix Factorization
How can this be used?
Given Decompose into and

Then solve for And then
solve for
31
Matrix Factorization
How is this better or faster than Gauss
Elimination? Lets look at computational time. n
number of equations
To decompose A, time is proportional to To
solve and time is
proportional to
32
Matrix Factorization
Therefore, total computational time for LU
Decomposition is proportional to
or
Gauss Elimination computation time is
proportional to
How is this better?
33
Matrix Factorization
What about a situation where the C vector
changes?
In VU factorization, VU decomposition of A is
independent of the C vector, therefore it only
needs to be done once. Let m the number of
times the C vector changes The computational
times are proportional to
VU factorization Gauss
Elimination
Consider a 100 equation set with 50
right hand side vectors VU factorization
Gauss Elimination
34
Matrix Factorization
Another Advantage Finding the Inverse of a Matrix
VU Factorization Gauss Elimination
For large values of n
35
Matrix Factorization
Method Decompose A to V and U
U is the same as the coefficient matrix at the
end of the forward elimination step. V is
obtained using the multipliers that were used in
the forward elimination process
36
Matrix Factorization
Finding the U matrix Using the Forward
Elimination Procedure of Gauss Elimination
37
Matrix Factorization
Finding the U matrix Using the Forward
Elimination Procedure of Gauss Elimination
38
Matrix Factorization
Finding the V matrix Using the multipliers used
during the Forward Elimination Procedure
From the first step of forward elimination
From the second step of forward elimination
39
Matrix Factorization
Does ?
40
Matrix Factorization
Example Solving simultaneous linear equations
using VU factorization
Solve the following set of linear equations using
VU Factorization
Using the procedure for finding the V and U
matrices
41
Matrix Factorization
Example Solving simultaneous linear equations
using VU factorization
Complete the forward substitution to solve for
42
Matrix Factorization
Example Solving simultaneous linear equations
using VU factorization
Set Solve for
The 3 equations become
43
Matrix Factorization
Example Solving simultaneous linear equations
using VU factorization
From the 3rd equation
Substituting in a3 and using the second equation
44
Matrix Factorization
Example Solving simultaneous linear equations
using VU factorization
Substituting in a3 and a2 using the first equation
Hence the Solution Vector is
45
Gauss Method
An iterative method.
  • Basic Procedure
  • Algebraically solve each linear equation for xi
  • Assume an initial guess solution array
  • Solve for each xi and repeat
  • Use absolute relative approximate error after
    each iteration to check if error is within a
    prespecified tolerance.

46
Gauss Method
Why?
The Gauss Method allows the user to control
round-off error. Elimination methods such as
Gaussian Elimination and VU Factorization are
prone to round-off error. Also If the physics
of the problem is understood, a close initial
guess can be made, decreasing the number of
iterations needed.
47
Gauss Method
Algorithm
A set of n equations and n unknowns
If the diagonal elements are non-zero Rewrite
each equation solving for the corresponding
unknown ex First equation, solve for x1 Second
equation, solve for x2
. . .
. . .
48
Gauss Method
Algorithm
Rewriting each equation
From Equation 1 From equation 2 From equation
n-1 From equation n
49
Gauss Method
Algorithm
General Form of each equation
50
Gauss Method
Gauss Algorithm
General Form for any row i
How or where can this equation be used?
51
Gauss Method
Solve for the unknowns
Use rewritten equations to solve for each value
of xi. Important Remember to use the most recent
value of xi. Which means to apply values
calculated to the calculations remaining in the
current iteration.
Assume an initial guess for X
52
Gauss Method
Calculate the Absolute Relative Approximate Error
So when has the answer been found? The
iterations are stopped when the absolute relative
approximate error is less than a pre-specified
tolerance for all unknowns.
53
Gauss-Seidel Method
  • Improved method

54
Gauss-Seidel Method Pitfall
Diagonally dominant The coefficient on the
diagonal must be at least equal to the sum of the
other coefficients in that row and at least one
row with a diagonal coefficient greater than the
sum of the other coefficients in that row.
Which coefficient matrix is diagonally dominant?
Most physical systems do result in simultaneous
linear equations that have diagonally dominant
coefficient matrices.
55
Gauss-Seidel Method Example
Given the system of equations
The coefficient matrix is


With an initial guess of
Will the solution converge using the Gauss-Seidel
method?
56
Gauss-Seidel Method Example
Checking if the coefficient matrix is diagonally
dominant

The inequalities are all true and at least one
row is strictly greater than Therefore The
solution should converge using the Gauss-Seidel
Method
57
Gauss-Siedel Method Example
Rewriting each equation
With an initial guess of





58
Gauss-Siedel Method Example
The absolute relative approximate error




The maximum absolute relative error after the
first iteration is 100
59
Gauss-Siedel Method Example
After Iteration 1

After Iteration 2
Substituting the x values into the equations

60
Gauss-Siedel Method Example
Iteration 2 absolute relative approximate error


The maximum absolute relative error after the
first iteration is 240.62 This is much larger
than the maximum absolute relative error obtained
in iteration 1. Is this a problem?
61
Gauss-Siedel Method Example
Repeating more iterations, the following values
are obtained
Iteration a1 a2 a3
1 2 3 4 5 6 0.50000 0.14679 0.74275 0.94675 0.99177 0.99919 67.662 240.62 80.23 21.547 4.5394 0.74260 4.900 3.7153 3.1644 3.0281 3.0034 3.0001 100.00 31.887 17.409 4.5012 0.82240 0.11000 3.0923 3.8118 3.9708 3.9971 4.0001 4.0001 67.662 18.876 4.0042 0.65798 0.07499 0.00000
The solution obtained is close to the
exact solution of
62
Comparison of different methods
  • Gauss is slow convergent however, more stable.
  • Gauss-Seidel might not be stable however, when
    stable, converges fast.
  • Hotelling and bodewig uses an improved iterative
    method for matrix inversion.
  • In Relaxation techniques small corrections will
    be applied to each element at each iteration.

63
Transformations and Eigenvlues
  • Many problems of the form
  • Can be written as
  • Where S is a diagonal Matrix.
  • Where the primed vectors represent the original
    vectors in the new space.

64
Eigenvalues and Eigenvectors
Write a Comment
User Comments (0)
About PowerShow.com