Parallel Implementations of Gaussian Elimination
|
|
- Thomas Goodwin
- 6 years ago
- Views:
Transcription
1 s of Western Michigan University January 27, 2012 CS 6260: in Parallel
2 Linear systems of equations General form of a linear system of equations is given by a 11 x a 1n x n = b 1 a 21 x a 2n x n = b 2.. a m1 x a mn x n = b m where a ij s and b i s are known and we are solving for x i s. CS 6260: in Parallel
3 Ax = b More compactly, we can rewrite system of linear equations in the form Ax = b where A = a 11 a a 1n a 21 a a 2n a m1 a m2... a mn, x = x 1 x 2. x n, b = b 1 b 2. b m CS 6260: in Parallel
4 Why study solutions of Ax = b 1 One of the most fundamental problems in the field of scientific computing 2 Arises in many applications: CS 6260: in Parallel
5 Why study solutions of Ax = b 1 One of the most fundamental problems in the field of scientific computing 2 Arises in many applications: Chemical engineering Interpolation Structural analysis Regression Analysis Numerical ODEs and PDEs CS 6260: in Parallel
6 Methods for solving Ax = b 1 Direct methods 2 Iterative methods CS 6260: in Parallel
7 Methods for solving Ax = b 1 Direct methods obtain the exact solution (in real arithmetic) in finitely many operations 2 Iterative methods generate sequence of approximations that converge in the limit to the solution CS 6260: in Parallel
8 Methods for solving Ax = b 1 Direct methods obtain the exact solution (in real arithmetic) in finitely many operations Gaussian elimination (LU factorization) QR factorization WZ factorization 2 Iterative methods generate sequence of approximations that converge in the limit to the solution Jacobi iteration Gauss-Seidal iteration SOR method (successive over-relaxation) CS 6260: in Parallel
9 Methods for solving Ax = b 1 Direct methods obtain the exact solution (in real arithmetic) in finitely many operations Gaussian elimination (LU factorization) QR factorization WZ factorization 2 Iterative methods generate sequence of approximations that converge in the limit to the solution Jacobi iteration Gauss-Seidal iteration SOR method (successive over-relaxation) CS 6260: in Parallel
10 Motivation When solving Ax = b we will assume throughout this presentation that A is non-singular and A and b are known a 11 x a 1n x n = b 1 a 21 x a 2n x n = b 2.. a n1 x a nn x n = b n. CS 6260: in Parallel
11 Motivation Assuming that a 11 =0wefirstsubtracta 21 /a 11 times the first equation from the second equation to eliminate the coefficient x 1 in the second equation, and so on until the coefficients of x 1 in the last n 1 rows have all been eliminated. This gives the modified system of equations CS 6260: in Parallel
12 Motivation Assuming that a 11 =0wefirstsubtracta 21 /a 11 times the first equation from the second equation to eliminate the coefficient x 1 in the second equation, and so on until the coefficients of x 1 in the last n 1 rows have all been eliminated. This gives the modified system of equations where a (1) ij a 11 x 1 + a 12 x a 1n x n = b 1 a (1) 22 x a (1) 2n x n = b (1) 2... a (1) n1 x a nn (1) x n = b n (1), a i1 = a ij a 1j b (1) a i1 i = b i b 1 i, j =1, 2,...,n. a 11 a 11 CS 6260: in Parallel
13 (Forward Reduction) Applying the same process the last n 1 equations of the modified system to eliminate coefficients of x 2 in the last n 2 equations, and so on, until the entire system has been reduced to the (upper) triangular form a 11 a 12 a 1n a (1) 22 a (1) 2n.... a nn (n 1) x 1 x 2 x n = b 1 b (1) 2. b (n 1) n. CS 6260: in Parallel
14 (Forward Reduction) Applying the same process the last n 1 equations of the modified system to eliminate coefficients of x 2 in the last n 2 equations, and so on, until the entire system has been reduced to the (upper) triangular form a 11 a 12 a 1n a (1) 22 a (1) 2n.... a nn (n 1) x 1 x 2 x n = b 1 b (1) 2. b (n 1) n. The superscripts indicate the number of times the elements had to be changed. CS 6260: in Parallel
15 Example Perform the forward reduction the the system given below x x 2 = x 3 1 CS 6260: in Parallel
16 Example Perform the forward reduction the the system given below x x 2 = x 3 1 We start by writing down the augmented matrix for the given system: CS 6260: in Parallel
17 Example We perform appropriate row operations to obtain: CS 6260: in Parallel
18 Example We perform appropriate row operations to obtain: R 2 ( 2 4)R 1 R 2 R 3 ( 1 4 )R 1 R CS 6260: in Parallel
19 Example We perform appropriate row operations to obtain: R 2 ( 2 4)R 1 R 2 R 3 ( 1 4 )R 1 R 3 R 3 ( )R 2 R CS 6260: in Parallel
20 Example Note that the row operations used to eliminate x 1 from the second and the third equations are equivalent to multiplying on the left the augmented matrix: = CS 6260: in Parallel
21 Example Note that the row operations used to eliminate x 1 from the second and the third equations are equivalent to multiplying on the left the augmented matrix: L = CS 6260: in Parallel
22 Example Note that the row operations used to eliminate x 1 from the second and the third equations are equivalent to multiplying on the left the augmented matrix: L L 1 [A b] = CS 6260: in Parallel
23 Example L L 1 L = L2 L 1 Thus, we can write [A b] = (L 2 L 1 ) A = L A = U CS 6260: in Parallel
24 LU factorization Motivation In general, we can think of row operations as multiplying with matrices on the left, thus the i th elimination step is equivalent as multiplication on the left by 1... L i = 1 l i+1,i..... l n,i 1 CS 6260: in Parallel
25 LU factorization Motivation Continuing in this fashion we obtain LAx = Lb, L = Ln 1 L 2 L 1, CS 6260: in Parallel
26 LU factorization Motivation Continuing in this fashion we obtain LAx = Lb, L = Ln 1 L 2 L 1, A = L 1 U = LU. CS 6260: in Parallel
27 LU factorization Motivation Continuing in this fashion we obtain A = L 1 U = LU. Thus, the Gaussian elimination algorithm for solving Ax = b is mathematically equivalent to the three-step process: 1. Factor A = LU 2. Solve (forward substitution) Ly = b 3. Solve (back substitution) Ux = y. CS 6260: in Parallel
28 LU factorization Motivation Continuing in this fashion we obtain A = L 1 U = LU. Thus, the Gaussian elimination algorithm for solving Ax = b is mathematically equivalent to the three-step process: 1. Factor A = LU 2. Solve (forward substitution) Ly = b 3. Solve (back substitution) Ux = y. Order of operations: n n2 n 3 multiplications/divisions n n2 2 5n 6. CS 6260: in Parallel
29 Need for partial pivoting Apply LU factorization without pivoting to A = 1 1 in three-decimal-digit floating point arithmetic. CS 6260: in Parallel
30 Need for partial pivoting A = Solution: L and U are easily obtainable and are given by: 1 0 L = fl(1/10 4, fl(1/10 4 ) rounds to 10 4, ) U = 0 fl(1 10 4, fl( ) rounds to ) so LU = = but A =. 1 1 CS 6260: in Parallel
31 Need for partial pivoting A = but LU = Remark 1: Note that original a 22 has been entirely lost from the computation by subtracting 10 4 from it. Thus if we were to use this LU factorization in order to solve a system there would be no way to guarantee an accurate answer. This is called numerical instability. CS 6260: in Parallel
32 Need for partial pivoting A = but LU = Remark 1: Note that original a 22 has been entirely lost from the computation by subtracting 10 4 from it. Thus if we were to use this LU factorization in order to solve a system there would be no way to guarantee an accurate answer. This is called numerical instability. Remark 2: Suppose we attempted to solve system Ax =[1, 2] T for x using this LU decomposition. The correct answer is x [1, 1] T.Instead,ifweweretostickwiththe three-digit-floating point arithmetic we would get an answer x =[0, 1] T which is completely erroneous. CS 6260: in Parallel
33 Motivation CS 6260: in Parallel
34 Row-Oriented Algorithm Pipelined Implementation Row-Oriented Algorithm 1 Determination of the local pivot element, 2 Determination of the global pivot element, 3 Exchange of the pivot row, 4 Distribution of the pivot row, 5 Computation of the elimination factors, 6 Computation of the matrix elements. CS 6260: in Parallel
35 Row-Oriented Algorithm Row-Oriented Algorithm Pipelined Implementation Remark 1: The computation of the solution vector x in the backward substitution is inherently serial, since the values of x k depend on each other and are computed one after another. Thus, in step k the processor P q owning row k computers the value of x k and sends the value to all other processors by a single broadcast operation. CS 6260: in Parallel
36 Row-Oriented Algorithm Row-Oriented Algorithm Pipelined Implementation Remark 1: The computation of the solution vector x in the backward substitution is inherently serial, since the values of x k depend on each other and are computed one after another. Thus, in step k the processor P q owning row k computers the value of x k and sends the value to all other processors by a single broadcast operation. Remark 2: Note that the data distribution is not quite adequate. For example, suppose we are at the i th step and that i > m n/p where m is a natural number. In that case, processors p 0, p 1,...,p m 1 are idle since all their work is done until the back substitution step at the very end. Hence there is an issue with load balancing. CS 6260: in Parallel
37 Row-Oriented Algorithm Row-Oriented Algorithm Pipelined Implementation Remark 2: Note that the data distribution is not quite adequate. For example, suppose we are at the i th step and that i > m n/p where m is a natural number. In that case, processors p 0, p 1,...,p m 1 are idle since all their work is done until the back substitution step at the very end. Hence there is an issue with load balancing. block-cyclic row distribution CS 6260: in Parallel
38 Row-Oriented Algorithm Row-Oriented Algorithm Pipelined Implementation Remark 3: We could also consider column-oriented data distribution. In that case, at the k th step the processor that owns k th column would have all needed data to compute the new pivot. On the one hand, this would reduce communication between processors that we had when considering row-oriented data distribution. On the other hand, with column orientation pivot determination is all done serially. Thus, in case when n >> p (which is often case) row oriented data distribution might be more advantageous since the pivot determination is done in parallel CS 6260: in Parallel
39 Pipelined Implementation Row-Oriented Algorithm Pipelined Implementation 1 Challenges and possible approaches Send ahead strategy CS 6260: in Parallel
40 Pipelined Implementation Row-Oriented Algorithm Pipelined Implementation 1 Challenges and possible approaches Send ahead strategy Challenges with pivoting CS 6260: in Parallel
41 Pipelined Implementation Row-Oriented Algorithm Pipelined Implementation 1 Challenges and possible approaches Send ahead strategy Challenges with pivoting Column pivoting CS 6260: in Parallel
42 Pipelined Implementation Row-Oriented Algorithm Pipelined Implementation 1 Challenges and possible approaches Send ahead strategy Challenges with pivoting Column pivoting 2 Advantages It allows asynchronous execution: processes can reduce their portions of the augmented matrix as soon as the pivot rows are available CS 6260: in Parallel
43 Pipelined Implementation Row-Oriented Algorithm Pipelined Implementation 1 Challenges and possible approaches Send ahead strategy Challenges with pivoting Column pivoting 2 Advantages It allows asynchronous execution: processes can reduce their portions of the augmented matrix as soon as the pivot rows are available It allows processes to overlap communication and computation times. CS 6260: in Parallel
44 Final Remarks Compute communication and computation times. CS 6260: in Parallel
45 Final Remarks Compute communication and computation times. Largely depends on what kind of distributed system used (usual configurations that have been studying are 1D and 2D meshes as well as hypercube). CS 6260: in Parallel
46 Final Remarks Compute communication and computation times. Largely depends on what kind of distributed system used (usual configurations that have been studying are 1D and 2D meshes as well as hypercube). Compute isoefficiency functions and determine scalability. CS 6260: in Parallel
47 Final Remarks Compute communication and computation times. Largely depends on what kind of distributed system used (usual configurations that have been studying are 1D and 2D meshes as well as hypercube). Compute isoefficiency functions and determine scalability. Row (column) oriented algorithms, stable but not scalable. Pipelined implementation is perfectly scalable. Structured systems Symmetric/Hermitian. Banded and upper (lower) triangular. Positive definite (CG method). CS 6260: in Parallel
48 Final Remarks Compute communication and computation times. Largely depends on what kind of distributed system used (usual configurations that have been studying are 1D and 2D meshes as well as hypercube). Compute isoefficiency functions and determine scalability. Row (column) oriented algorithms, stable but not scalable. Pipelined implementation is perfectly scalable. Structured systems Symmetric/Hermitian. Banded and upper (lower) triangular. Positive definite (CG method). Sparse systems and iterative methods (numerical PDEs) CS 6260: in Parallel
Dense Matrix Algorithms
Dense Matrix Algorithms Ananth Grama, Anshul Gupta, George Karypis, and Vipin Kumar To accompany the text Introduction to Parallel Computing, Addison Wesley, 2003. Topic Overview Matrix-Vector Multiplication
More informationLecture 12 (Last): Parallel Algorithms for Solving a System of Linear Equations. Reference: Introduction to Parallel Computing Chapter 8.
CZ4102 High Performance Computing Lecture 12 (Last): Parallel Algorithms for Solving a System of Linear Equations - Dr Tay Seng Chuan Reference: Introduction to Parallel Computing Chapter 8. 1 Topic Overview
More informationNumerical Linear Algebra
Numerical Linear Algebra Probably the simplest kind of problem. Occurs in many contexts, often as part of larger problem. Symbolic manipulation packages can do linear algebra "analytically" (e.g. Mathematica,
More informationIterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms
Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear
More informationNumerical Algorithms
Chapter 10 Slide 464 Numerical Algorithms Slide 465 Numerical Algorithms In textbook do: Matrix multiplication Solving a system of linear equations Slide 466 Matrices A Review An n m matrix Column a 0,0
More informationNumerical Methods 5633
Numerical Methods 5633 Lecture 7 Marina Krstic Marinkovic mmarina@maths.tcd.ie School of Mathematics Trinity College Dublin Marina Krstic Marinkovic 1 / 10 5633-Numerical Methods Organisational To appear
More informationSection 3.1 Gaussian Elimination Method (GEM) Key terms
Section 3.1 Gaussian Elimination Method (GEM) Key terms Rectangular systems Consistent system & Inconsistent systems Rank Types of solution sets RREF Upper triangular form & back substitution Nonsingular
More informationLecture 4: Principles of Parallel Algorithm Design (part 4)
Lecture 4: Principles of Parallel Algorithm Design (part 4) 1 Mapping Technique for Load Balancing Minimize execution time Reduce overheads of execution Sources of overheads: Inter-process interaction
More informationMatrix multiplication
Matrix multiplication Standard serial algorithm: procedure MAT_VECT (A, x, y) begin for i := 0 to n - 1 do begin y[i] := 0 for j := 0 to n - 1 do y[i] := y[i] + A[i, j] * x [j] end end MAT_VECT Complexity:
More informationLARP / 2018 ACK : 1. Linear Algebra and Its Applications - Gilbert Strang 2. Autar Kaw, Transforming Numerical Methods Education for STEM Graduates
Triangular Factors and Row Exchanges LARP / 28 ACK :. Linear Algebra and Its Applications - Gilbert Strang 2. Autar Kaw, Transforming Numerical Methods Education for STEM Graduates Then there were three
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 20: Sparse Linear Systems; Direct Methods vs. Iterative Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 26
More informationChapter 8 Dense Matrix Algorithms
Chapter 8 Dense Matrix Algorithms (Selected slides & additional slides) A. Grama, A. Gupta, G. Karypis, and V. Kumar To accompany the text Introduction to arallel Computing, Addison Wesley, 23. Topic Overview
More informationOutline. Parallel Algorithms for Linear Algebra. Number of Processors and Problem Size. Speedup and Efficiency
1 2 Parallel Algorithms for Linear Algebra Richard P. Brent Computer Sciences Laboratory Australian National University Outline Basic concepts Parallel architectures Practical design issues Programming
More informationCE 601: Numerical Methods Lecture 5. Course Coordinator: Dr. Suresh A. Kartha, Associate Professor, Department of Civil Engineering, IIT Guwahati.
CE 601: Numerical Methods Lecture 5 Course Coordinator: Dr. Suresh A. Kartha, Associate Professor, Department of Civil Engineering, IIT Guwahati. Elimination Methods For a system [A]{x} = {b} where [A]
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 1 Don t you just invert
More informationParallelizing LU Factorization
Parallelizing LU Factorization Scott Ricketts December 3, 2006 Abstract Systems of linear equations can be represented by matrix equations of the form A x = b LU Factorization is a method for solving systems
More informationHigh Performance Computing Programming Paradigms and Scalability Part 6: Examples of Parallel Algorithms
High Performance Computing Programming Paradigms and Scalability Part 6: Examples of Parallel Algorithms PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering (CiE) Scientific Computing
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert
More informationContents. F10: Parallel Sparse Matrix Computations. Parallel algorithms for sparse systems Ax = b. Discretized domain a metal sheet
Contents 2 F10: Parallel Sparse Matrix Computations Figures mainly from Kumar et. al. Introduction to Parallel Computing, 1st ed Chap. 11 Bo Kågström et al (RG, EE, MR) 2011-05-10 Sparse matrices and storage
More informationAddition/Subtraction flops. ... k k + 1, n (n k)(n k) (n k)(n + 1 k) n 1 n, n (1)(1) (1)(2)
1 CHAPTER 10 101 The flop counts for LU decomposition can be determined in a similar fashion as was done for Gauss elimination The major difference is that the elimination is only implemented for the left-hand
More informationCS 6210 Fall 2016 Bei Wang. Review Lecture What have we learnt in Scientific Computing?
CS 6210 Fall 2016 Bei Wang Review Lecture What have we learnt in Scientific Computing? Let s recall the scientific computing pipeline observed phenomenon mathematical model discretization solution algorithm
More informationExam Design and Analysis of Algorithms for Parallel Computer Systems 9 15 at ÖP3
UMEÅ UNIVERSITET Institutionen för datavetenskap Lars Karlsson, Bo Kågström och Mikael Rännar Design and Analysis of Algorithms for Parallel Computer Systems VT2009 June 2, 2009 Exam Design and Analysis
More informationSolving Sparse Linear Systems. Forward and backward substitution for solving lower or upper triangular systems
AMSC 6 /CMSC 76 Advanced Linear Numerical Analysis Fall 7 Direct Solution of Sparse Linear Systems and Eigenproblems Dianne P. O Leary c 7 Solving Sparse Linear Systems Assumed background: Gauss elimination
More informationIterative Methods for Linear Systems
Iterative Methods for Linear Systems 1 the method of Jacobi derivation of the formulas cost and convergence of the algorithm a Julia function 2 Gauss-Seidel Relaxation an iterative method for solving linear
More informationChemical Engineering 541
Chemical Engineering 541 Computer Aided Design Methods Direct Solution of Linear Systems 1 Outline 2 Gauss Elimination Pivoting Scaling Cost LU Decomposition Thomas Algorithm (Iterative Improvement) Overview
More informationx = 12 x = 12 1x = 16
2.2 - The Inverse of a Matrix We've seen how to add matrices, multiply them by scalars, subtract them, and multiply one matrix by another. The question naturally arises: Can we divide one matrix by another?
More informationComputational Methods CMSC/AMSC/MAPL 460. Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science
Computational Methods CMSC/AMSC/MAPL 460 Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Zero elements of first column below 1 st row multiplying 1 st
More informationMATH 423 Linear Algebra II Lecture 17: Reduced row echelon form (continued). Determinant of a matrix.
MATH 423 Linear Algebra II Lecture 17: Reduced row echelon form (continued). Determinant of a matrix. Row echelon form A matrix is said to be in the row echelon form if the leading entries shift to the
More informationIterative Sparse Triangular Solves for Preconditioning
Euro-Par 2015, Vienna Aug 24-28, 2015 Iterative Sparse Triangular Solves for Preconditioning Hartwig Anzt, Edmond Chow and Jack Dongarra Incomplete Factorization Preconditioning Incomplete LU factorizations
More informationAutar Kaw Benjamin Rigsby. Transforming Numerical Methods Education for STEM Undergraduates
Autar Kaw Benjamin Rigsy http://nmmathforcollegecom Transforming Numerical Methods Education for STEM Undergraduates http://nmmathforcollegecom LU Decomposition is another method to solve a set of simultaneous
More informationLinear Equation Systems Iterative Methods
Linear Equation Systems Iterative Methods Content Iterative Methods Jacobi Iterative Method Gauss Seidel Iterative Method Iterative Methods Iterative methods are those that produce a sequence of successive
More informationLecture 17: Array Algorithms
Lecture 17: Array Algorithms CS178: Programming Parallel and Distributed Systems April 4, 2001 Steven P. Reiss I. Overview A. We talking about constructing parallel programs 1. Last time we discussed sorting
More informationIntroduction to Parallel. Programming
University of Nizhni Novgorod Faculty of Computational Mathematics & Cybernetics Introduction to Parallel Section 9. Programming Parallel Methods for Solving Linear Systems Gergel V.P., Professor, D.Sc.,
More information1 2 (3 + x 3) x 2 = 1 3 (3 + x 1 2x 3 ) 1. 3 ( 1 x 2) (3 + x(0) 3 ) = 1 2 (3 + 0) = 3. 2 (3 + x(0) 1 2x (0) ( ) = 1 ( 1 x(0) 2 ) = 1 3 ) = 1 3
6 Iterative Solvers Lab Objective: Many real-world problems of the form Ax = b have tens of thousands of parameters Solving such systems with Gaussian elimination or matrix factorizations could require
More information10/26/ Solving Systems of Linear Equations Using Matrices. Objectives. Matrices
6.1 Solving Systems of Linear Equations Using Matrices Objectives Write the augmented matrix for a linear system. Perform matrix row operations. Use matrices and Gaussian elimination to solve systems.
More informationA Study of Numerical Methods for Simultaneous Equations
A Study of Numerical Methods for Simultaneous Equations Er. Chandan Krishna Mukherjee B.Sc.Engg., ME, MBA Asstt. Prof. ( Mechanical ), SSBT s College of Engg. & Tech., Jalgaon, Maharashtra Abstract: -
More informationCS321 Introduction To Numerical Methods
CS3 Introduction To Numerical Methods Fuhua (Frank) Cheng Department of Computer Science University of Kentucky Lexington KY 456-46 - - Table of Contents Errors and Number Representations 3 Error Types
More informationECE 204 Numerical Methods for Computer Engineers MIDTERM EXAMINATION /4:30-6:00
ECE 4 Numerical Methods for Computer Engineers ECE 4 Numerical Methods for Computer Engineers MIDTERM EXAMINATION --7/4:-6: The eamination is out of marks. Instructions: No aides. Write your name and student
More informationSparse Linear Systems
1 Sparse Linear Systems Rob H. Bisseling Mathematical Institute, Utrecht University Course Introduction Scientific Computing February 22, 2018 2 Outline Iterative solution methods 3 A perfect bipartite
More informationAim. Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity. Structure and matrix sparsity: Overview
Aim Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk What should a 2-hour PhD lecture on structure
More informationChapter 14: Matrix Iterative Methods
Chapter 14: Matrix Iterative Methods 14.1INTRODUCTION AND OBJECTIVES This chapter discusses how to solve linear systems of equations using iterative methods and it may be skipped on a first reading of
More informationCS 770G - Parallel Algorithms in Scientific Computing
CS 770G - Parallel lgorithms in Scientific Computing Dense Matrix Computation II: Solving inear Systems May 28, 2001 ecture 6 References Introduction to Parallel Computing Kumar, Grama, Gupta, Karypis,
More informationContents. Preface xvii Acknowledgments. CHAPTER 1 Introduction to Parallel Computing 1. CHAPTER 2 Parallel Programming Platforms 11
Preface xvii Acknowledgments xix CHAPTER 1 Introduction to Parallel Computing 1 1.1 Motivating Parallelism 2 1.1.1 The Computational Power Argument from Transistors to FLOPS 2 1.1.2 The Memory/Disk Speed
More informationSparse matrices, graphs, and tree elimination
Logistics Week 6: Friday, Oct 2 1. I will be out of town next Tuesday, October 6, and so will not have office hours on that day. I will be around on Monday, except during the SCAN seminar (1:25-2:15);
More information2. Use elementary row operations to rewrite the augmented matrix in a simpler form (i.e., one whose solutions are easy to find).
Section. Gaussian Elimination Our main focus in this section is on a detailed discussion of a method for solving systems of equations. In the last section, we saw that the general procedure for solving
More informationSeptember, a 11 x 1 +a 12 x a 1n x n =b 1 a 21 x 1 +a 22 x a 2n x n =b 2.. (1) a n 1 x 1 +a n2 x a nn x n = b n.
September, 1998 PHY307F/407F - Computational Physics Background Material for the Exercise - Solving Systems of Linear Equations David Harrison This document discusses techniques to solve systems of linear
More informationUppsala University Department of Information technology. Hands-on 1: Ill-conditioning = x 2
Uppsala University Department of Information technology Hands-on : Ill-conditioning Exercise (Ill-conditioned linear systems) Definition A system of linear equations is said to be ill-conditioned when
More informationPrinciple Of Parallel Algorithm Design (cont.) Alexandre David B2-206
Principle Of Parallel Algorithm Design (cont.) Alexandre David B2-206 1 Today Characteristics of Tasks and Interactions (3.3). Mapping Techniques for Load Balancing (3.4). Methods for Containing Interaction
More informationLINPACK Benchmark. on the Fujitsu AP The LINPACK Benchmark. Assumptions. A popular benchmark for floating-point performance. Richard P.
1 2 The LINPACK Benchmark on the Fujitsu AP 1000 Richard P. Brent Computer Sciences Laboratory The LINPACK Benchmark A popular benchmark for floating-point performance. Involves the solution of a nonsingular
More informationContents. I The Basic Framework for Stationary Problems 1
page v Preface xiii I The Basic Framework for Stationary Problems 1 1 Some model PDEs 3 1.1 Laplace s equation; elliptic BVPs... 3 1.1.1 Physical experiments modeled by Laplace s equation... 5 1.2 Other
More informationChapter LU Decomposition More Examples Electrical Engineering
Chapter 4.7 LU Decomposition More Examples Electrical Engineering Example Three-phase loads e common in AC systems. When the system is balanced the analysis can be simplified to a single equivalent rcuit
More informationNumerical Analysis I - Final Exam Matrikelnummer:
Dr. Behrens Center for Mathematical Sciences Technische Universität München Winter Term 2005/2006 Name: Numerical Analysis I - Final Exam Matrikelnummer: I agree to the publication of the results of this
More informationLinear systems of equations
Linear systems of equations Michael Quinn Parallel Programming with MPI and OpenMP material do autor Terminology Back substitution Gaussian elimination Outline Problem System of linear equations Solve
More informationAssignment 1 (concept): Solutions
CS10b Data Structures and Algorithms Due: Thursday, January 0th Assignment 1 (concept): Solutions Note, throughout Exercises 1 to 4, n denotes the input size of a problem. 1. (10%) Rank the following functions
More informationAMS527: Numerical Analysis II
AMS527: Numerical Analysis II A Brief Overview of Finite Element Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao SUNY Stony Brook AMS527: Numerical Analysis II 1 / 25 Overview Basic concepts Mathematical
More informationMidterm Exam with solutions
Midterm Exam with solutions CS227-Introduction to Scientific Computation November 8, 2011 1. The following is a transcript of a MATLAB session. >> x=1/62.5 x = 0.016000000000000 >> y=(1+x)-1 y = 0.016000000000000
More informationProject Report. 1 Abstract. 2 Algorithms. 2.1 Gaussian elimination without partial pivoting. 2.2 Gaussian elimination with partial pivoting
Project Report Bernardo A. Gonzalez Torres beaugonz@ucsc.edu Abstract The final term project consist of two parts: a Fortran implementation of a linear algebra solver and a Python implementation of a run
More informationComputational Methods CMSC/AMSC/MAPL 460. Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science
Computational Methods CMSC/AMSC/MAPL 460 Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Some special matrices Matlab code How many operations and memory
More informationCS Elementary Graph Algorithms & Transform-and-Conquer
CS483-10 Elementary Graph Algorithms & Transform-and-Conquer Outline Instructor: Fei Li Room 443 ST II Office hours: Tue. & Thur. 1:30pm - 2:30pm or by appointments Depth-first Search cont Topological
More informationThe Simplex Algorithm. Chapter 5. Decision Procedures. An Algorithmic Point of View. Revision 1.0
The Simplex Algorithm Chapter 5 Decision Procedures An Algorithmic Point of View D.Kroening O.Strichman Revision 1.0 Outline 1 Gaussian Elimination 2 Satisfiability with Simplex 3 General Simplex Form
More informationSection 1.5. Finding Linear Equations
Section 1.5 Finding Linear Equations Using Slope and a Point to Find an Equation of a Line Example Find an equation of a line that has slope m = 3 and contains the point (2, 5). Solution Substitute m =
More informationEFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI
EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI 1 Akshay N. Panajwar, 2 Prof.M.A.Shah Department of Computer Science and Engineering, Walchand College of Engineering,
More informationNumerical Methods for PDEs : Video 9: 2D Finite Difference February 14, Equations / 29
22.520 Numerical Methods for PDEs Video 9 2D Finite Difference Equations February 4, 205 22.520 Numerical Methods for PDEs Video 9 2D Finite Difference February 4, Equations 205 / 29 Thought Experiment
More informationDM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini
DM545 Linear and Integer Programming Lecture 2 The Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 3. 4. Standard Form Basic Feasible Solutions
More informationTHE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECH- NOLOGY ITERATIVE LINEAR SOLVERS
ITERATIVE LIEAR SOLVERS. Objectives The goals of the laboratory workshop are as follows: to learn basic properties of iterative methods for solving linear least squares problems, to study the properties
More informationProf. Dr. Stefan Funken, Prof. Dr. Alexander Keller, Prof. Dr. Karsten Urban 11. Januar Scientific Computing Parallele Algorithmen
Prof. Dr. Stefan Funken, Prof. Dr. Alexander Keller, Prof. Dr. Karsten Urban 11. Januar 2007 Scientific Computing Parallele Algorithmen Page 2 Scientific Computing 11. Januar 2007 Funken / Keller / Urban
More informationAlgorithms and Data Structures
Charles A. Wuethrich Bauhaus-University Weimar - CogVis/MMC June 22, 2017 1/51 Introduction Matrix based Transitive hull All shortest paths Gaussian elimination Random numbers Interpolation and Approximation
More informationBlocking SEND/RECEIVE
Message Passing Blocking SEND/RECEIVE : couple data transfer and synchronization - Sender and receiver rendezvous to exchange data P P SrcP... x : =... SEND(x, DestP)... DestP... RECEIVE(y,SrcP)... M F
More informationParallel Threshold-based ILU Factorization
A short version of this paper appears in Supercomputing 997 Parallel Threshold-based ILU Factorization George Karypis and Vipin Kumar University of Minnesota, Department of Computer Science / Army HPC
More informationSolving Systems Using Row Operations 1 Name
The three usual methods of solving a system of equations are graphing, elimination, and substitution. While these methods are excellent, they can be difficult to use when dealing with three or more variables.
More informationSolutions for Operations Research Final Exam
Solutions for Operations Research Final Exam. (a) The buffer stock is B = i a i = a + a + a + a + a + a 6 + a 7 = + + + + + + =. And the transportation tableau corresponding to the transshipment problem
More informationPARALLEL METHODS FOR SOLVING PARTIAL DIFFERENTIAL EQUATIONS. Ioana Chiorean
5 Kragujevac J. Math. 25 (2003) 5 18. PARALLEL METHODS FOR SOLVING PARTIAL DIFFERENTIAL EQUATIONS Ioana Chiorean Babeş-Bolyai University, Department of Mathematics, Cluj-Napoca, Romania (Received May 28,
More informationCS6015 / LARP ACK : Linear Algebra and Its Applications - Gilbert Strang
Solving and CS6015 / LARP 2018 ACK : Linear Algebra and Its Applications - Gilbert Strang Introduction Chapter 1 concentrated on square invertible matrices. There was one solution to Ax = b and it was
More informationMatrix Inverse 2 ( 2) 1 = 2 1 2
Name: Matrix Inverse For Scalars, we have what is called a multiplicative identity. This means that if we have a scalar number, call it r, then r multiplied by the multiplicative identity equals r. Without
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 21 Outline 1 Course
More informationNow, start over with a smarter method. But it is still heavily relied on the power of matlab. But we can see the pattern of the matrix elements.
Instructions for starting on Project 5. Warning: many of the instructions are not that smart or efficient. It is part of a learning process. Start by introducing some data points for x and for y: octave:>
More informationCOMPUTER SCIENCE 314 Numerical Methods SPRING 2013 ASSIGNMENT # 2 (25 points) January 22
COMPUTER SCIENCE 314 Numerical Methods SPRING 2013 ASSIGNMENT # 2 (25 points) January 22 Announcements Office hours: Instructor Teaching Assistant Monday 4:00 5:00 Tuesday 2:30 3:00 4:00 5:00 Wednesday
More informationSOLVING SYSTEMS OF LINEAR INTERVAL EQUATIONS USING THE INTERVAL EXTENDED ZERO METHOD AND MULTIMEDIA EXTENSIONS
Please cite this article as: Mariusz Pilarek, Solving systems of linear interval equations using the "interval extended zero" method and multimedia extensions, Scientific Research of the Institute of Mathematics
More informationEXTENSION. a 1 b 1 c 1 d 1. Rows l a 2 b 2 c 2 d 2. a 3 x b 3 y c 3 z d 3. This system can be written in an abbreviated form as
EXTENSION Using Matrix Row Operations to Solve Systems The elimination method used to solve systems introduced in the previous section can be streamlined into a systematic method by using matrices (singular:
More information1 Exercise: 1-D heat conduction with finite elements
1 Exercise: 1-D heat conduction with finite elements Reading This finite element example is based on Hughes (2000, sec. 1.1-1.15. 1.1 Implementation of the 1-D heat equation example In the previous two
More informationDesign of Parallel Algorithms. Models of Parallel Computation
+ Design of Parallel Algorithms Models of Parallel Computation + Chapter Overview: Algorithms and Concurrency n Introduction to Parallel Algorithms n Tasks and Decomposition n Processes and Mapping n Processes
More informationCSCE 5160 Parallel Processing. CSCE 5160 Parallel Processing
HW #9 10., 10.3, 10.7 Due April 17 { } Review Completing Graph Algorithms Maximal Independent Set Johnson s shortest path algorithm using adjacency lists Q= V; for all v in Q l[v] = infinity; l[s] = 0;
More informationAMSC/CMSC 460 Final Exam, Fall 2007
AMSC/CMSC 460 Final Exam, Fall 2007 Show all work. You may leave arithmetic expressions in any form that a calculator could evaluate. By putting your name on this paper, you agree to abide by the university
More informationIdentity Matrix: >> eye(3) ans = Matrix of Ones: >> ones(2,3) ans =
Very Basic MATLAB Peter J. Olver January, 2009 Matrices: Type your matrix as follows: Use space or, to separate entries, and ; or return after each row. >> [;5 0-3 6;; - 5 ] or >> [,5,6,-9;5,0,-3,6;7,8,5,0;-,,5,]
More informationMathematical preliminaries and error analysis
Mathematical preliminaries and error analysis Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan August 28, 2011 Outline 1 Round-off errors and computer arithmetic IEEE
More informationParallel Computing. Slides credit: M. Quinn book (chapter 3 slides), A Grama book (chapter 3 slides)
Parallel Computing 2012 Slides credit: M. Quinn book (chapter 3 slides), A Grama book (chapter 3 slides) Parallel Algorithm Design Outline Computational Model Design Methodology Partitioning Communication
More information3. Replace any row by the sum of that row and a constant multiple of any other row.
Math Section. Section.: Solving Systems of Linear Equations Using Matrices As you may recall from College Algebra or Section., you can solve a system of linear equations in two variables easily by applying
More informationNatural Quartic Spline
Natural Quartic Spline Rafael E Banchs INTRODUCTION This report describes the natural quartic spline algorithm developed for the enhanced solution of the Time Harmonic Field Electric Logging problem As
More informationFor example, the system. 22 may be represented by the augmented matrix
Matrix Solutions to Linear Systems A matrix is a rectangular array of elements. o An array is a systematic arrangement of numbers or symbols in rows and columns. Matrices (the plural of matrix) may be
More informationPerformance Evaluation of Multiple and Mixed Precision Iterative Refinement Method and its Application to High-Order Implicit Runge-Kutta Method
Performance Evaluation of Multiple and Mixed Precision Iterative Refinement Method and its Application to High-Order Implicit Runge-Kutta Method Tomonori Kouya Shizuoa Institute of Science and Technology,
More informationLinear Arrays. Chapter 7
Linear Arrays Chapter 7 1. Basics for the linear array computational model. a. A diagram for this model is P 1 P 2 P 3... P k b. It is the simplest of all models that allow some form of communication between
More informationSolving Algebraic Equations
Lesson 4. Solving Algebraic Equations 3 3 3 3 3 8 8 4 Add 3 to both sides. Divide both sides by. 4 gives the solution of the equation 3. Check: Substitute 4 for x into the original equation. 3 4 3 When
More informationF. THOMSON LEIGHTON INTRODUCTION TO PARALLEL ALGORITHMS AND ARCHITECTURES: ARRAYS TREES HYPERCUBES
F. THOMSON LEIGHTON INTRODUCTION TO PARALLEL ALGORITHMS AND ARCHITECTURES: ARRAYS TREES HYPERCUBES MORGAN KAUFMANN PUBLISHERS SAN MATEO, CALIFORNIA Contents Preface Organization of the Material Teaching
More informationParallel Numerical Algorithms
Parallel Numerical Algorithms Chapter 5 Vector and Matrix Products Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign CS 554 / CSE 512 Michael T. Heath Parallel
More informationFast Iterative Solvers for Markov Chains, with Application to Google's PageRank. Hans De Sterck
Fast Iterative Solvers for Markov Chains, with Application to Google's PageRank Hans De Sterck Department of Applied Mathematics University of Waterloo, Ontario, Canada joint work with Steve McCormick,
More informationChapter 4: Implicit Error Detection
4. Chpter 5 Chapter 4: Implicit Error Detection Contents 4.1 Introduction... 4-2 4.2 Network error correction... 4-2 4.3 Implicit error detection... 4-3 4.4 Mathematical model... 4-6 4.5 Simulation setup
More informationNumerical Method (2068 Third Batch)
1. Define the types of error in numerical calculation. Derive the formula for secant method and illustrate the method by figure. There are different types of error in numerical calculation. Some of them
More informationProcess model formulation and solution, 3E4 Computer software tutorial - Tutorial 3
Process model formulation and solution, 3E4 Computer software tutorial - Tutorial 3 Kevin Dunn, dunnkg@mcmaster.ca October 2 Tutorial solutions: Elliot Cameron. Tutorial objectives Understand computer
More informationContents. Implementing the QR factorization The algebraic eigenvalue problem. Applied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationJournal of Engineering Research and Studies E-ISSN
Journal of Engineering Research and Studies E-ISS 0976-79 Research Article SPECTRAL SOLUTIO OF STEADY STATE CODUCTIO I ARBITRARY QUADRILATERAL DOMAIS Alavani Chitra R 1*, Joshi Pallavi A 1, S Pavitran
More information