THE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECH- NOLOGY ITERATIVE LINEAR SOLVERS
|
|
- Bernard Gibson
- 5 years ago
- Views:
Transcription
1 ITERATIVE LIEAR SOLVERS. Objectives The goals of the laboratory workshop are as follows: to learn basic properties of iterative methods for solving linear least squares problems, to study the properties (convergence, computational complexity, sensitivity to errors) of the selected linear solvers, to train the skills in coding the selected algorithms in Matlab and in using the Matlab functions for solving systems of linear equations. The workshop is scheduled for 3 academic hours.. Introduction * Iterative linear solvers attempt to iteratively approximate the solution = [ x ] I the system of linear equations: M where = [ ] I a ij A is a coefficient matrix and [ ] M following updates: x to Ax = b, () b = b I is a data vector, using the k ( ) ( k+ ) ( x = f x ), A, b ( ) where x k is the approximation to the solution x in the k-th iterative step, and f (,, is the ( update function determined by the underlying iterative method, where k ) * lim f x, A, b x. i, k j ) ( ) This laboratory workshop focuses on the two main classes of iterative solvers: stationary iterative methods (Richardson, Landweber, ewton, Jacobi, Gauss-Seidel, SOR) and Krylov subspace methods (CG, BiCG, BiCGSTAB, GMRES, QMR, LSQR). 3. Preparation. The expected time needed for the preparation to this workshop is 9 hours. 3.. Reading []. A. Bjorck, umerical Methods for Least Squares Problems, SIAM, Philadelphia, 996, []. G. Golub, C. F. Van Loan, Matrix Computations, The John Hopkins University Press, (Third Edition), 996. [3]. Y. Saad, Iterative Methods for Sparse Linear Systems, SIAM, 003. [4]. J. ocedal, S. J. Wright, umerical Optimization, Springer, 999, pp. 63, Project co-financed by European Union within European Social Fund
2 [5]. C. D. Meyer, Matrix Analysis and Applied Linear Algebra, SIAM, 000, [6]. Ch. Zarowski, An Introduction to umerical Analysis for Electrical and Computer Engineers, Wiley, 004, 3.. Problems At the beginning of the laboratory workshop each student should know the answers to the following questions: How to check whether a given system of linear equations has none, one, or many solutions? What is an over-determined, determined, and under-determined system of linear equations? How to formulate a Least Square (LS) problem to a system of linear equations? How to derive a system of normal equations to an LS problem? What is a probabilistic and a geometric interpretation of LS approximations (draw an example of curve fitting)? How to estimate the coefficients of a linear combination of the basis functions to fit data in the LS sense (linear regression problem)? Give examples of the fundamental stationary iterative methods for solving systems of linear equations. Give examples of the Krylov subspace methods for solving systems of linear equations. How to define the spectral radius? Explain how the CG method works (give an example). What is a convergence rate between the steepest descent gradient method and the ewton method? What is the Hessian matrix and how to calculate it for the linear LS problem? What is preconditioning and how does it work? What is the difference between parallel and coordinate-wise implementation of iterative methods? How to solve the constrained linear LS problems? 3.3. Detailed preparation Each group of students ( 3 persons) is expected to accomplish the following tasks:. code the selected algorithms for solving systems of linear equations in Matlab,. solve the selected problems with the coded algorithms, 3. compare the results (convergence rate, elapsed time, computational cost, etc.) obtained with the coded algorithms and with in-built functions in Matlab, 4. draw the conclusions. Project co-financed by European Union within European Social Fund
3 Problems to be solved The following problems that are given or can be expressed in terms of systems of linear equations should be solved with the coded algorithms. Problem : Solve the following system with the selected iterative solvers u v= 0 u + v w = 0. v + w z = 0 w+ z = 5 Estimate the computational costs and convergence rates. Start the iterations from zero-value initial guess. Problem : Solve the following system of linear equations using the selected iterative solvers: x+ x + x3 = x+ x + x3 =. x + x + x3 = Compare and discuss the computational costs with respect to convergence rates. Start the iterations from x (0) = 0. Problem 3: Estimate the exact solution to the following system of linear equations using the Gaussian elimination: 0.835x x = 0.68, 0.333x+ 0.66x = 0.067, Then slightly perturb b from to and compute iterative approximations to the perturbed system using the selected iterative solvers. Draw the distance between the exact solution and the iterative approximations versus iterations. Start the iterations from (0) (0) x = x = 0. Estimate the computational costs for each tested methods and compare them with the costs for the Gaussian elimination. Problem 4: Solve the system of linear equations: Ax = b, where: = A, = [ ] b K T, Project co-financed by European Union within European Social Fund 3
4 with the selected iterative solvers and compare the results to the Gaussian elimination. Start the iterations from (0) ( k) ( ) x = 0. Draw the residual error: r = b-ax k and the approximation error: x -x, where the exact solution is obtained with the Gaussian elimination. Use any preconditioning and compare the results. Problem 5: Let A = a ij R, with [ ] T b = K R. Solve the system Ax = b a = ij i + j (Hilbert matrix), and for = 5, 0 and 0, using the selected it- (0) erative methods. Start the iterations from x = 0. Draw the residual error: ( k) ( k) r = b-ax. Which iterative solver and preconditioning are the most appropriate for solving this problem? Problem 6: For = 5, 0, 5, 0 solve the system of linear equations: Ax = b, where sin() sin() sin( ) T b = K R, and A is the Vandermonde matrix with the (-)- T th column equal to [ K ]. To generate this matrix, use the function vander(.) from Matlab. Then perturb the vector b with a zero-mean uncorrelated Gaussian error (SR = 30 db) and determine the approximation error versus the size. Problem 7: Let [ 0 ] T c = K R and = [ 0 + ] * is the first column and r is the first row of the Toeplitz, and [ ] r K R. Assuming c T x = K R, perform the forward projection: Ax * = b, and solve the system of linear equations with the selected iterative solvers. Draw the errors: r = b-ax and x -x. Which ( k) ( k) method gives the best approximations? Explain why the approximation error is large even for a large number of iterations. Problem 8: Let 0 O 0 O A = 0 O O 0 R, b [ K ] T = R Solve the system Ax = b with the selected iterative solvers for = 0, 00, 000, 0000, For each draw the residual error ( k r ) = b-ax ( k) versus the k-the iterative step. Explain the difference in the convergence behavior. T Problem 9: Let Ax = b, where A= I C C, the symbol denotes the Kronecker product, I M M R is an identity matrix, C R is a random matrix with a uniform distribution, M = 00, and = 50, and x ( 0, I M ). Find the iterative method that solves the above system of linear equations with the lowest computational cost. Estimate the cost with a Project co-financed by European Union within European Social Fund 4
5 roughly calculated number of flops and with the elapsed time. Draw the errors x -x ver- sus iterations. 0 0 Problem 0: Solve Ax = b, where A R is the Hilbert matrix, and x ( 0, I 0 ), with the selected iterative solvers and with the Gaussian elimination in Matlab (x := A\b). Then make a small change in an entry of A or b, and compare the solutions. Draw the errors x -x versus iterations for the perturbed system. Algorithms to be coded Students are expected to code the following algorithms and apply them to the abovementioned nonlinear problems. Algorithm : The Richardson s first-order or Landweber method: [], Chapter 7, Section 7..3., pp Algorithm : The Jacobi s method: [] Chapter 7, Section 7..3., pp ; [] Chapter 0, Section 0..-., pp. 50 5; [3] Chapter 4, Section 4.., pp Algorithm 3: The Gauss-Seidel method: [] Chapter 7, Section 7..3., pp ; [] Chapter 0, Section , pp ; [3] Chapter 4, Section 4.., pp Algorithm 4: The Successive Over-Relaxation (SOR) method: [] Chapter 7, Section 7..4., pp ; [] Chapter 0, Section , pp ; [3] Chapter 4, Section 4.., pp Algorithm 5: The Steepest Descent (SD) method: [] Chapter 0, Section 0..-., pp. 50 5; [3] Chapter 5, Section 5.3., pp. 3 33, Algorithm 5.. Algorithm 6: The Conjugate Gradient (CG) method: [] Chapter 0, Section 0.., pp , Algorithm 0...; [3] Chapter 6, Section 6.7, pp. 74 8, Algorithm 6.7.; [4] Chapter 5, pp. 0 33, Algorithm 5.. Algorithm 7: The Preconditioned Conjugate Gradient (PCG) method: [] Chapter 0, Section 0.3., pp , Algorithm 0.3..; [3] Chapter 9, Section 9., pp , Algorithm 9..; [4] Chapter 5, pp. 0 33, Algorithm 5.3. Algorithm 8: The CGLS method: [] Chapter 7, Section 7.4., pp , Algorithm Algorithm 9: The PCCGLS method: [] Chapter 7, Section 7.4., pp , Algorithm Algorithm 0: The LSQR method: [] Chapter 7, Section 7.6.3, pp , Algorithm Project co-financed by European Union within European Social Fund 5
6 Algorithm : The CGR method: [] Chapter 0, Section 0.4., pp , Algorithm 0.4..; [3] Chapter 8, Section 8.3, pp , Algorithm 8.4. Algorithm : The CGE method: [] Chapter 0, Section 0.4., pp , Algorithm 0.4..; [3] Chapter 8, Section 8.3, pp , Algorithm 8.5. Algorithm 3: The BiCG method: [3] Chapter 7, Section 7.3, pp. 09, Algorithm 7.3. Algorithm 4: The BiCGSTAB method: [3] Chapter 7, Section 7.4, pp. 6 9, Algorithm 7.6. Algorithm 5: The CGS method: [3] Chapter 7, Section 7.4, pp. 3 6, Algorithm 7.8. Algorithm 6: The GMRES method: [] Chapter 0, Section 0.4.4, pp , Algorithm 0.4.4; [3] Chapter 6, Section 6.5, pp. 57 7, Algorithm 6.9 (standard version), Algorithm 6.0 (Householder version), Algorithm 6. (restarted version). Algorithm 7: The QMR method: [] Chapter 0, Section 0.4.7, p ; [3] Chapter 7, Section 7.3, pp. 3, Algorithm 7.4. Algorithm 8: The Kaczmarz method: [5] Chapter 5, Section 5.3.9, pp Algorithm 9: The Oblique Projection method: [5] Chapter 5, Section 5.3.0, pp Algorithm 0: The ewton method: [6], Section 8.4, pp Content of report The report should contain: introductory page, detailed mathematical description of the analyzed problems, a basic description of the coded algorithms, the Matlab code (together with the detailed end-line comments) of the analyzed algorithm, discussion on convergence properties of the analyzed methods, the results obtained with the coded algorithms (the residual error versus iterations: ( k ) * b-ax ; approximation error (if possible): x -x where x is the exact solution; estimation of computational costs), the results obtained with the Matlab in-built functions, conclusions Project co-financed by European Union within European Social Fund 6
AMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 20: Sparse Linear Systems; Direct Methods vs. Iterative Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 26
More informationIterative Methods for Solving Linear Problems
Iterative Methods for Solving Linear Problems When problems become too large (too many data points, too many model parameters), SVD and related approaches become impractical. Iterative Methods for Solving
More informationAlgebraic Iterative Methods for Computed Tomography
Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic
More informationContents. I The Basic Framework for Stationary Problems 1
page v Preface xiii I The Basic Framework for Stationary Problems 1 1 Some model PDEs 3 1.1 Laplace s equation; elliptic BVPs... 3 1.1.1 Physical experiments modeled by Laplace s equation... 5 1.2 Other
More informationIterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms
Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear
More informationAlgebraic Iterative Methods for Computed Tomography
Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic
More informationAN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS
AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS Seyed Abolfazl Shahzadehfazeli 1, Zainab Haji Abootorabi,3 1 Parallel Processing Laboratory, Yazd University,
More informationParallel Numerical Algorithms
Parallel Numerical Algorithms Chapter 4 Sparse Linear Systems Section 4.3 Iterative Methods Michael T. Heath and Edgar Solomonik Department of Computer Science University of Illinois at Urbana-Champaign
More informationSELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND
Student Submission for the 5 th OpenFOAM User Conference 2017, Wiesbaden - Germany: SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND TESSA UROIĆ Faculty of Mechanical Engineering and Naval Architecture, Ivana
More informationPreliminary Investigations on Resilient Parallel Numerical Linear Algebra Solvers
SIAM EX14 Workshop July 7, Chicago - IL reliminary Investigations on Resilient arallel Numerical Linear Algebra Solvers HieACS Inria roject Joint Inria-CERFACS lab INRIA Bordeaux Sud-Ouest Luc Giraud joint
More informationNAG Fortran Library Routine Document F11DSF.1
NAG Fortran Library Routine Document Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent
More informationnag sparse nsym sol (f11dec)
f11 Sparse Linear Algebra f11dec nag sparse nsym sol (f11dec) 1. Purpose nag sparse nsym sol (f11dec) solves a real sparse nonsymmetric system of linear equations, represented in coordinate storage format,
More informationTHE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECHNOLOGY METAHEURISTICS
METAHEURISTICS 1. Objectives The goals of the laboratory workshop are as follows: to learn basic properties of evolutionary computation techniques and other metaheuristics for solving various global optimization
More informationCS 6210 Fall 2016 Bei Wang. Review Lecture What have we learnt in Scientific Computing?
CS 6210 Fall 2016 Bei Wang Review Lecture What have we learnt in Scientific Computing? Let s recall the scientific computing pipeline observed phenomenon mathematical model discretization solution algorithm
More informationPARALLEL METHODS FOR SOLVING PARTIAL DIFFERENTIAL EQUATIONS. Ioana Chiorean
5 Kragujevac J. Math. 25 (2003) 5 18. PARALLEL METHODS FOR SOLVING PARTIAL DIFFERENTIAL EQUATIONS Ioana Chiorean Babeş-Bolyai University, Department of Mathematics, Cluj-Napoca, Romania (Received May 28,
More information1 2 (3 + x 3) x 2 = 1 3 (3 + x 1 2x 3 ) 1. 3 ( 1 x 2) (3 + x(0) 3 ) = 1 2 (3 + 0) = 3. 2 (3 + x(0) 1 2x (0) ( ) = 1 ( 1 x(0) 2 ) = 1 3 ) = 1 3
6 Iterative Solvers Lab Objective: Many real-world problems of the form Ax = b have tens of thousands of parameters Solving such systems with Gaussian elimination or matrix factorizations could require
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert
More information1 Exercise: 1-D heat conduction with finite elements
1 Exercise: 1-D heat conduction with finite elements Reading This finite element example is based on Hughes (2000, sec. 1.1-1.15. 1.1 Implementation of the 1-D heat equation example In the previous two
More informationParallel Implementations of Gaussian Elimination
s of Western Michigan University vasilije.perovic@wmich.edu January 27, 2012 CS 6260: in Parallel Linear systems of equations General form of a linear system of equations is given by a 11 x 1 + + a 1n
More informationUppsala University Department of Information technology. Hands-on 1: Ill-conditioning = x 2
Uppsala University Department of Information technology Hands-on : Ill-conditioning Exercise (Ill-conditioned linear systems) Definition A system of linear equations is said to be ill-conditioned when
More informationXinyu Dou Acoustics Technology Center, Motorola, Inc., Schaumburg, Illinois 60196
A unified boundary element method for the analysis of sound and shell-like structure interactions. II. Efficient solution techniques Shaohai Chen and Yijun Liu a) Department of Mechanical Engineering,
More informationNAG Library Function Document nag_sparse_nsym_sol (f11dec)
f11 Large Scale Linear Systems NAG Library Function Document nag_sparse_nsym_sol () 1 Purpose nag_sparse_nsym_sol () solves a real sparse nonsymmetric system of linear equations, represented in coordinate
More informationIntroduction to Parallel Computing
Introduction to Parallel Computing W. P. Petersen Seminar for Applied Mathematics Department of Mathematics, ETHZ, Zurich wpp@math. ethz.ch P. Arbenz Institute for Scientific Computing Department Informatik,
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 1 Don t you just invert
More informationCS 542G: Solving Sparse Linear Systems
CS 542G: Solving Sparse Linear Systems Robert Bridson November 26, 2008 1 Direct Methods We have already derived several methods for solving a linear system, say Ax = b, or the related leastsquares problem
More informationEfficient Use of Iterative Solvers in Nested Topology Optimization
Efficient Use of Iterative Solvers in Nested Topology Optimization Oded Amir, Mathias Stolpe and Ole Sigmund Technical University of Denmark Department of Mathematics Department of Mechanical Engineering
More informationContents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.
page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5
More informationSpeeding up SOR Solvers for Constraint-based GUIs with a Warm-Start Strategy
Speeding up SOR Solvers for Constraint-based GUIs with a Warm-Start Strategy Noreen Jamil, Johannes Müller, Christof Lutteroth and Gerald Weber Department of Computer Science University of Auckland Private
More informationNonsymmetric Problems. Abstract. The eect of a threshold variant TPABLO of the permutation
Threshold Ordering for Preconditioning Nonsymmetric Problems Michele Benzi 1, Hwajeong Choi 2, Daniel B. Szyld 2? 1 CERFACS, 42 Ave. G. Coriolis, 31057 Toulouse Cedex, France (benzi@cerfacs.fr) 2 Department
More informationCpu time [s] BICGSTAB_RPC CGS_RPC BICGSTAB_LPC BICG_RPC BICG_LPC LU_NPC
Application of Non-stationary Iterative Methods to an Exact Newton-Raphson Solution Process for Power Flow Equations Rainer Bacher, Eric Bullinger Swiss Federal Institute of Technology (ETH), CH-809 Zurich,
More informationNumerical Methods to Solve 2-D and 3-D Elliptic Partial Differential Equations Using Matlab on the Cluster maya
Numerical Methods to Solve 2-D and 3-D Elliptic Partial Differential Equations Using Matlab on the Cluster maya David Stonko, Samuel Khuvis, and Matthias K. Gobbert (gobbert@umbc.edu) Department of Mathematics
More informationWhitepaper: Software-Defined Events (SDEs) in MAGMA-Sparse
Whitepaper: Software-Defined Events (SDEs) in MAGMA-Sparse ECP PEEKS: Production-ready, Exascale-Enabled, Krylov Solvers for Exascale Computing Heike Jagode Anthony Danalis Hartwig Anzt Ichitaro Yamazaki
More informationSolving Sparse Linear Systems. Forward and backward substitution for solving lower or upper triangular systems
AMSC 6 /CMSC 76 Advanced Linear Numerical Analysis Fall 7 Direct Solution of Sparse Linear Systems and Eigenproblems Dianne P. O Leary c 7 Solving Sparse Linear Systems Assumed background: Gauss elimination
More informationDeveloping a High Performance Software Library with MPI and CUDA for Matrix Computations
Developing a High Performance Software Library with MPI and CUDA for Matrix Computations Bogdan Oancea 1, Tudorel Andrei 2 1 Nicolae Titulescu University of Bucharest, e-mail: bogdanoancea@univnt.ro, Calea
More informationECE 204 Numerical Methods for Computer Engineers MIDTERM EXAMINATION /4:30-6:00
ECE 4 Numerical Methods for Computer Engineers ECE 4 Numerical Methods for Computer Engineers MIDTERM EXAMINATION --7/4:-6: The eamination is out of marks. Instructions: No aides. Write your name and student
More informationFINITE ELEMENT SOLUTION OF NAVIER-STOKES EQUATIONS USING KRYLOV SUBSPACE METHODS
HEFAT2014 10 th International Conference on Heat Transfer, Fluid Mechanics and Thermodynamics 14 16 July 2014 Orlando, Florida FINITE ELEMENT SOLUTION OF NAVIER-STOKES EQUATIONS USING KRYLOV SUBSPACE METHODS
More informationHYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE
HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER AVISHA DHISLE PRERIT RODNEY ADHISLE PRODNEY 15618: PARALLEL COMPUTER ARCHITECTURE PROF. BRYANT PROF. KAYVON LET S
More informationCS 395T Lecture 12: Feature Matching and Bundle Adjustment. Qixing Huang October 10 st 2018
CS 395T Lecture 12: Feature Matching and Bundle Adjustment Qixing Huang October 10 st 2018 Lecture Overview Dense Feature Correspondences Bundle Adjustment in Structure-from-Motion Image Matching Algorithm
More informationPolymath 6. Overview
Polymath 6 Overview Main Polymath Menu LEQ: Linear Equations Solver. Enter (in matrix form) and solve a new system of simultaneous linear equations. NLE: Nonlinear Equations Solver. Enter and solve a new
More informationAccelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method
Accelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method Wenyong Pan 1, Kris Innanen 1 and Wenyuan Liao 2 1. CREWES Project, Department of Geoscience,
More informationComputing with vectors and matrices in C++
CS319: Scientific Computing (with C++) Computing with vectors and matrices in C++ Week 7: 9am and 4pm, 22 Feb 2017 1 Introduction 2 Solving linear systems 3 Jacobi s Method 4 Implementation 5 Vectors 6
More informationMathematical Methods and Computational Algorithms for Complex Networks. Benard Abola
Mathematical Methods and Computational Algorithms for Complex Networks Benard Abola Division of Applied Mathematics, Mälardalen University Department of Mathematics, Makerere University Second Network
More informationHigh Performance Computing Programming Paradigms and Scalability Part 6: Examples of Parallel Algorithms
High Performance Computing Programming Paradigms and Scalability Part 6: Examples of Parallel Algorithms PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering (CiE) Scientific Computing
More informationA hybrid GMRES and TV-norm based method for image restoration
A hybrid GMRES and TV-norm based method for image restoration D. Calvetti a, B. Lewis b and L. Reichel c a Department of Mathematics, Case Western Reserve University, Cleveland, OH 44106 b Rocketcalc,
More informationPerformance Strategies for Parallel Mathematical Libraries Based on Historical Knowledgebase
Performance Strategies for Parallel Mathematical Libraries Based on Historical Knowledgebase CScADS workshop 29 Eduardo Cesar, Anna Morajko, Ihab Salawdeh Universitat Autònoma de Barcelona Objective Mathematical
More informationLecture 15: More Iterative Ideas
Lecture 15: More Iterative Ideas David Bindel 15 Mar 2010 Logistics HW 2 due! Some notes on HW 2. Where we are / where we re going More iterative ideas. Intro to HW 3. More HW 2 notes See solution code!
More informationMultigrid Methods for Markov Chains
Multigrid Methods for Markov Chains Hans De Sterck Department of Applied Mathematics, University of Waterloo collaborators Killian Miller Department of Applied Mathematics, University of Waterloo, Canada
More informationGTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013
GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS Kyle Spagnoli Research Engineer @ EM Photonics 3/20/2013 INTRODUCTION» Sparse systems» Iterative solvers» High level benchmarks»
More informationComparison of different solvers for two-dimensional steady heat conduction equation ME 412 Project 2
Comparison of different solvers for two-dimensional steady heat conduction equation ME 412 Project 2 Jingwei Zhu March 19, 2014 Instructor: Surya Pratap Vanka 1 Project Description The purpose of this
More informationLinear Equation Systems Iterative Methods
Linear Equation Systems Iterative Methods Content Iterative Methods Jacobi Iterative Method Gauss Seidel Iterative Method Iterative Methods Iterative methods are those that produce a sequence of successive
More informationLeast-Squares Fitting Algorithms of the NIST Algorithm Testing System
[J. Res. Natl. Inst. Stand. Technol. 103, 633 (1998)] Least-Squares Fitting Algorithms of the NIST Algorithm Testing System Volume 103 Number 6 November December 1998 Craig M. Shakarji National Institute
More informationAn Introduction to Numerical Analysis
Weimin Han AMCS & Dept of Math University of Iowa MATH:38 Example 1 Question: What is the area of the region between y = e x 2 and the x-axis for x 1? Answer: Z 1 e x 2 dx = : 1.9.8.7.6.5.4.3.2.1 1.5.5
More informationEFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI
EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI 1 Akshay N. Panajwar, 2 Prof.M.A.Shah Department of Computer Science and Engineering, Walchand College of Engineering,
More informationIMPROVING THE PERFORMANCE OF THE LINEAR SYSTEMS SOLVERS USING CUDA
2036 Challenges of the Knowledge Society. IT IMPROVING THE PERFORMANCE OF THE LINEAR SYSTEMS SOLVERS USING CUDA BOGDAN OANCEA * TUDOREL ANDREI ** RALUCA MARIANA DRAGOESCU *** Abstract Parallel computing
More informationOn the regularizing power of multigrid-type algorithms. Marco Donatelli. Stefano Serra Capizzano
On the regularizing power of multigrid-type algorithms Marco Donatelli Stefano Serra Capizzano Università dell Insubria, Dip. Fisica e Matematica - Sede di Como Via Valleggio 11-22100 Como, Italy (donatelli@uninsubria.it)
More informationPerformance of Implicit Solver Strategies on GPUs
9. LS-DYNA Forum, Bamberg 2010 IT / Performance Performance of Implicit Solver Strategies on GPUs Prof. Dr. Uli Göhner DYNAmore GmbH Stuttgart, Germany Abstract: The increasing power of GPUs can be used
More informationAlgorithms, System and Data Centre Optimisation for Energy Efficient HPC
2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy
More informationAM205: lecture 2. 1 These have been shifted to MD 323 for the rest of the semester.
AM205: lecture 2 Luna and Gary will hold a Python tutorial on Wednesday in 60 Oxford Street, Room 330 Assignment 1 will be posted this week Chris will hold office hours on Thursday (1:30pm 3:30pm, Pierce
More informationExperimental Physics I & II "Junior Lab"
MIT OpenCourseWare http://ocw.mit.edu 8.13-14 Experimental Physics I & II "Junior Lab" Fall 2007 - Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationJournal of Engineering Research and Studies E-ISSN
Journal of Engineering Research and Studies E-ISS 0976-79 Research Article SPECTRAL SOLUTIO OF STEADY STATE CODUCTIO I ARBITRARY QUADRILATERAL DOMAIS Alavani Chitra R 1*, Joshi Pallavi A 1, S Pavitran
More informationSparse Linear Systems
1 Sparse Linear Systems Rob H. Bisseling Mathematical Institute, Utrecht University Course Introduction Scientific Computing February 22, 2018 2 Outline Iterative solution methods 3 A perfect bipartite
More information3-D Tomographic Reconstruction
Mitglied der Helmholtz-Gemeinschaft 3-D Tomographic Reconstruction of Atmospheric Trace Gas Concentrations for Infrared Limb-Imagers February 21, 2011, Nanocenter, USC Jörn Ungermann Tomographic trace
More informationMachine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013
Machine Learning Topic 5: Linear Discriminants Bryan Pardo, EECS 349 Machine Learning, 2013 Thanks to Mark Cartwright for his extensive contributions to these slides Thanks to Alpaydin, Bishop, and Duda/Hart/Stork
More informationLecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice
Stat60/CS94: Randomized Algorithms for Matrices and Data Lecture 11-10/09/013 Lecture 11: Randomized Least-squares Approximation in Practice Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these
More informationALF USER GUIDE. Date: September
ALF USER GUIDE R. VERFÜRTH Contents 1. Introduction 1 2. User interface 2 3. Domain and initial grid 3 4. Differential equation 9 5. Discretization 12 6. Solver 12 7. Mesh refinement 14 8. Number of nodes
More informationSolution of 2D Euler Equations and Application to Airfoil Design
WDS'6 Proceedings of Contributed Papers, Part I, 47 52, 26. ISBN 8-86732-84-3 MATFYZPRESS Solution of 2D Euler Equations and Application to Airfoil Design J. Šimák Charles University, Faculty of Mathematics
More informationUser Activity Recognition Based on Kalman Filtering Approach
User Activity Recognition Based on Kalman Filtering Approach EEC 592, Prosthesis Design and Control Final Project Report Gholamreza Khademi khademi.gh@gmail.com December 2, 214 Abstract Different control
More informationProf. Dr. Stefan Funken, Prof. Dr. Alexander Keller, Prof. Dr. Karsten Urban 11. Januar Scientific Computing Parallele Algorithmen
Prof. Dr. Stefan Funken, Prof. Dr. Alexander Keller, Prof. Dr. Karsten Urban 11. Januar 2007 Scientific Computing Parallele Algorithmen Page 2 Scientific Computing 11. Januar 2007 Funken / Keller / Urban
More informationIterative Methods for Linear Systems
Iterative Methods for Linear Systems 1 the method of Jacobi derivation of the formulas cost and convergence of the algorithm a Julia function 2 Gauss-Seidel Relaxation an iterative method for solving linear
More informationART Exhibit. Per Christian Hansen. Tommy Elfving Touraj Nikazad Hans Henrik B. Sørensen. joint work with, among others
ART Exhibit Per Christian Hansen joint work with, among others Tommy Elfving Touraj Nikazad Hans Henrik B. Sørensen About Me Forward problem Interests: numerical methods for inverse problems and tomography,
More informationA High-Order Accurate Unstructured GMRES Solver for Poisson s Equation
A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation Amir Nejat * and Carl Ollivier-Gooch Department of Mechanical Engineering, The University of British Columbia, BC V6T 1Z4, Canada
More informationSparse Matrices. This means that for increasing problem size the matrices become sparse and sparser. O. Rheinbach, TU Bergakademie Freiberg
Sparse Matrices Many matrices in computing only contain a very small percentage of nonzeros. Such matrices are called sparse ( dünn besetzt ). Often, an upper bound on the number of nonzeros in a row can
More informationGeneralized Principal Component Analysis CVPR 2007
Generalized Principal Component Analysis Tutorial @ CVPR 2007 Yi Ma ECE Department University of Illinois Urbana Champaign René Vidal Center for Imaging Science Institute for Computational Medicine Johns
More informationTwo-phase matrix splitting methods for asymmetric and symmetric LCP
Two-phase matrix splitting methods for asymmetric and symmetric LCP Daniel P. Robinson Department of Applied Mathematics and Statistics Johns Hopkins University Joint work with Feng, Nocedal, and Pang
More informationOptimization. Industrial AI Lab.
Optimization Industrial AI Lab. Optimization An important tool in 1) Engineering problem solving and 2) Decision science People optimize Nature optimizes 2 Optimization People optimize (source: http://nautil.us/blog/to-save-drowning-people-ask-yourself-what-would-light-do)
More informationReport of Linear Solver Implementation on GPU
Report of Linear Solver Implementation on GPU XIANG LI Abstract As the development of technology and the linear equation solver is used in many aspects such as smart grid, aviation and chemical engineering,
More informationParallel resolution of sparse linear systems by mixing direct and iterative methods
Parallel resolution of sparse linear systems by mixing direct and iterative methods Phyleas Meeting, Bordeaux J. Gaidamour, P. Hénon, J. Roman, Y. Saad LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix
More informationPreconditioning Linear Systems Arising from Graph Laplacians of Complex Networks
Preconditioning Linear Systems Arising from Graph Laplacians of Complex Networks Kevin Deweese 1 Erik Boman 2 1 Department of Computer Science University of California, Santa Barbara 2 Scalable Algorithms
More informationPerformance Studies for the Two-Dimensional Poisson Problem Discretized by Finite Differences
Performance Studies for the Two-Dimensional Poisson Problem Discretized by Finite Differences Jonas Schäfer Fachbereich für Mathematik und Naturwissenschaften, Universität Kassel Abstract In many areas,
More informationChapter 14: Matrix Iterative Methods
Chapter 14: Matrix Iterative Methods 14.1INTRODUCTION AND OBJECTIVES This chapter discusses how to solve linear systems of equations using iterative methods and it may be skipped on a first reading of
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Conjugate Direction Methods Barnabás Póczos & Ryan Tibshirani Conjugate Direction Methods 2 Books to Read David G. Luenberger, Yinyu Ye: Linear and Nonlinear Programming Nesterov:
More informationImprovements of the Discrete Dipole Approximation method
arxiv:physics/0006064v1 [physics.ao-ph] 26 Jun 2000 Improvements of the Discrete Dipole Approximation method Piotr J. Flatau Scripps Institution of Oceanography, University of California, San Diego, La
More informationPENALIZED LEAST-SQUARES IMAGE RECONSTRUCTION FOR BOREHOLE TOMOGRAPHY
Proceedings of ALGORITMY pp. 9 PENALIZED LEAST-SQUARES IMAGE RECONSTRUCTION FOR BOREHOLE TOMOGRAPHY CONSTANTIN POPA AND RAFAL ZDUNEK Abstract. The Algebraic Reconstruction Technique (ART), which is based
More informationOn Improving Computational Efficiency of Simplified Fluid Flow Models
1447 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 70, 2018 Guest Editors: Timothy G. Walmsley, Petar S. Varbanov, Rongxin Su, Jiří J. Klemeš Copyright 2018, AIDIC Servizi S.r.l. ISBN 978-88-95608-67-9;
More informationAdaptive Filters Algorithms (Part 2)
Adaptive Filters Algorithms (Part 2) Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Technology Digital Signal Processing and System
More informationMulti Layer Perceptron trained by Quasi Newton learning rule
Multi Layer Perceptron trained by Quasi Newton learning rule Feed-forward neural networks provide a general framework for representing nonlinear functional mappings between a set of input variables and
More informationStructure-preserving Smoothing for Seismic Amplitude Data by Anisotropic Diffusion using GPGPU
GPU Technology Conference 2016 April, 4-7 San Jose, CA, USA Structure-preserving Smoothing for Seismic Amplitude Data by Anisotropic Diffusion using GPGPU Joner Duarte jduartejr@tecgraf.puc-rio.br Outline
More informationA Lagrange method based L-curve for image restoration
Journal of Physics: Conference Series OPEN ACCESS A Lagrange method based L-curve for image restoration To cite this article: G Landi 2013 J. Phys.: Conf. Ser. 464 012011 View the article online for updates
More informationPROGRAMMING OF MULTIGRID METHODS
PROGRAMMING OF MULTIGRID METHODS LONG CHEN In this note, we explain the implementation detail of multigrid methods. We will use the approach by space decomposition and subspace correction method; see Chapter:
More informationRegression Solver. User Manual. Process Design and Gas Processing Laboratory Illinois Institute of Technology Chicago, IL,
Regression Solver User Manual Process Design and Gas Processing Laboratory Illinois Institute of Technology Chicago, IL, 60616. Copyright 2012-2016. All rights reserved. Introduction Regression Solver
More informationEmpirical Complexity of Laplacian Linear Solvers: Discussion
Empirical Complexity of Laplacian Linear Solvers: Discussion Erik Boman, Sandia National Labs Kevin Deweese, UC Santa Barbara John R. Gilbert, UC Santa Barbara 1 Simons Institute Workshop on Fast Algorithms
More informationConstraining strategies for Kaczmarz-like algorithms
Constraining strategies for Kaczmarz-like algorithms Ovidius University, Constanta, Romania Faculty of Mathematics and Computer Science Talk on April 7, 2008 University of Erlangen-Nürnberg The scanning
More informationParallel Hybrid Monte Carlo Algorithms for Matrix Computations
Parallel Hybrid Monte Carlo Algorithms for Matrix Computations V. Alexandrov 1, E. Atanassov 2, I. Dimov 2, S.Branford 1, A. Thandavan 1 and C. Weihrauch 1 1 Department of Computer Science, University
More informationImplicit schemes for wave models
Implicit schemes for wave models Mathieu Dutour Sikirić Rudjer Bo sković Institute, Croatia and Universität Rostock April 17, 2013 I. Wave models Stochastic wave modelling Oceanic models are using grids
More information6. Linear Discriminant Functions
6. Linear Discriminant Functions Linear Discriminant Functions Assumption: we know the proper forms for the discriminant functions, and use the samples to estimate the values of parameters of the classifier
More informationPerformance Analysis of Distributed Iterative Linear Solvers
Performance Analysis of Distributed Iterative Linear Solvers W.M. ZUBEREK and T.D.P. PERERA Department of Computer Science Memorial University St.John s, Canada A1B 3X5 Abstract: The solution of large,
More informationA parallel direct/iterative solver based on a Schur complement approach
A parallel direct/iterative solver based on a Schur complement approach Gene around the world at CERFACS Jérémie Gaidamour LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix project) February 29th, 2008
More informationNAG Library Function Document nag_sparse_sym_sol (f11jec)
f11 Large Scale Linear Systems f11jec NAG Library Function Document nag_sparse_sym_sol (f11jec) 1 Purpose nag_sparse_sym_sol (f11jec) solves a real sparse symmetric system of linear equations, represented
More informationIterative Sparse Triangular Solves for Preconditioning
Euro-Par 2015, Vienna Aug 24-28, 2015 Iterative Sparse Triangular Solves for Preconditioning Hartwig Anzt, Edmond Chow and Jack Dongarra Incomplete Factorization Preconditioning Incomplete LU factorizations
More informationFast Iterative Solvers for Markov Chains, with Application to Google's PageRank. Hans De Sterck
Fast Iterative Solvers for Markov Chains, with Application to Google's PageRank Hans De Sterck Department of Applied Mathematics University of Waterloo, Ontario, Canada joint work with Steve McCormick,
More information