Cpu time [s] BICGSTAB_RPC CGS_RPC BICGSTAB_LPC BICG_RPC BICG_LPC LU_NPC

Size: px
Start display at page:

Download "Cpu time [s] BICGSTAB_RPC CGS_RPC BICGSTAB_LPC BICG_RPC BICG_LPC LU_NPC"

Transcription

1 Application of Non-stationary Iterative Methods to an Exact Newton-Raphson Solution Process for Power Flow Equations Rainer Bacher, Eric Bullinger Swiss Federal Institute of Technology (ETH), CH-809 Zurich, Switzerland ABSTRACT The AC power ow is usually solved by the Newton- Raphson solution method. The main step is the linearization of the non-linear power ow equations and the subsequent solution of this linear system. The characteristics of this linear system of equations vary for dierent power ow implementations: Symmetric/unsymmetric, positive denite/indenite system matrices can result. Based on these characteristics dierent direct and iterative linear system solvers can be used to maximize performance and solution robustness. In this paper results are given of nonstationary iterative methods for unsymmetric, indefinite linear systems derived from power ow equations. Keywords Load ow analysis, Modelling and simulation, Non-stationary iterative methods 1 INTRODUCTION The power ow is a very well known algorithmic problem which is usually solved by the Newton- Raphson solution method. The linearization of the non-linear power ow equations yields a linear system to be solved by an appropriate linear system solver. All Newton-Raphson (NR) based power ow algorithms have in common that one large or two smaller size linear systems of equations must be solved during each Newton-Raphson iterative step. Beside the well known direct solution of a linear system of equations [1], the solution based on socalled non-stationary iterative methods has recently appeared in power applications. [] describes the rst application of the Conjugate Gradients (CG) method to the decoupled power ow. In [3] another application of CG methods applied to a static security power ow problem is described. For the fast decoupled power ow both papers state a signicant performance improvement of CG based methods compared to a direct solution. From all power ow approaches known today only the decoupled power ow satises the CG conditions of positive denite and symmetric linear system matrices without applying equation transformations. Practical usage of the conjugate gradient methods is only reached when preconditioning is applied to the linear system of equations. Good preconditioning will group the eigenvalues of the transformed linear system matrix together and will thus result in faster convergence. Again, there is a natural t between the preconditioned CG method and the decoupled power ow, since the decoupled power has constant linear system matrices. Thus, for every power ow this preconditioning matrix must be computed only once and remains constant for all Newton-Raphson steps. [4], [5] emphasize the fact that only good preconditioners allow an ecient implementation of CG methods to power ow equations. Several preconditioners such as ILU(0) (no ll-in factorization) or ILU(1) (approximate factorization with 1 neighbor ll-in term) together with block-bordered matrix permutation have been applied successfully to fast decoupled power ows with large networks. Recently, in [6], the rst application of nonstationary, iterative methods to the non-linear power ow problem has been described. A derivation of the \Krylov subspace power ow methodology" applied to the power ow problem is given to introduce power system application developers to the mathematical problem. The main distinction to the CG methods lies in the fact that the \Krylov subspace power ow methodology" is also applicable to unsymmetric, indenite linear system matrices. The so-called KSPF (Krylov Subspace Power Flow) which is derived in [6] does not need any explicit computation of the Jacobian terms during the iterations and power ow steps. Good convergence is shown with networks up to size 57 busses. In this paper implementation aspects of nonstationary methods with larger networks are discussed which are in principle applicable to all power ow formulations with unsymmetric and indenite linear system matrices. Mathematicians have de- 1

2 rived several methods to solve this type of linear system of equations: QMR (Quasi Minimal Residual), GMRES (Generalized Minimal Residual), BICG (Bi- Conjugate Gradients), CGS(Conjugate Gradients Squared), BICGSTAB (Bi-Conjugate Gradients Stabilized) are distinctly dierent methods for the solution of this class of problems. A good summary of these methods can be found in [7]. The GMRES-algorithm is the generalization of the CG algorithm for unsymmetric and indenite linear system matrices. Both algorithms have in common that the solution error (, see section 3) decreases from one iteration to the next and the exact solution is obtained within a given maximum number of iterations. This assumes exact numeric precision. All other non-stationary iterative methods use combinations of CG-like concepts and heuristics to obtain a solution for the linear system. As a consequence the of these methods is not guaranteed to decrease during the iterations. This paper gives details related to a successful computer implementation of non-stationary iterative methods applied to the non-linear power ow equations (1) and simulation results which should allow to get insight into the strengths and weaknesses of non-stationary iterative methods. After a problem denition in section, a special preconditioning of the linear system related to the Newton-Raphson solution process is given in section 3. In section 4, simulation results are discussed for various non-stationary methods applied to networks with 33 and 685 busses. The paper ends with conclusions in section 5 and the references. The direct solution based on factorization and a subsequent forward/backward substitution is not discussed in this paper, see [1]. Also, the mathematical theory for the non-stationary iterative method is not given in this paper due to space reasons, see [7]. An excellent reference to CG is [8]. The original description of CG is [9]. PROBLEM STATEMENT This paper deals with the iterative behavior of the solution of the linear system within the Newton- Raphson solution process for the non-linear power ow equations. The power ow equations for each node i are as follows: g i1 g i = e ip i +f i Q i e +f i i =,e iq i +f i P i e +f i i, P N j=1 (g ije j, b ij f j ) = 0 P N, j=1 (g ijf j + b ij e j ) = 0 i = 0 g i3 =,e i, f i + V (1) e i ;f i represent the real and imaginary part of nodal voltages, P i ;Q i represent the active and reactive nodal power, g ij ;b ij the real and imaginary part of the nodal admittance matrix Y ij, N = total number of nodes. For PQ-nodes the third equation g i3 is not needed. This power ow equations set (1) is called the \Current mismatch power ow" [10] since mainly nodal current equations are formulated. Note that the reactive powers Q i are unknown variables at PV generator nodes. These variables must be updated during all Newton-Raphson steps and linear system iterations 1. The Jacobian matrix terms are as follows: Jacobian diagonal block A ii 1 3 x = ei x = fi x = Qi ii, gii ii, bii ii + bii,ii, gii f i e i +f i, e i e i +f i,ei,fi 0 ii =, Pi (e i,f)+e i if i Q i (e i +f i ) ii =, Qi (f i,e i )+e if i P i (e i +f i ) (3) O-diagonal block A ij ;i 1 3 x=ej(i6=j) x=fj(i6=j) x=qj(i6=j),gij bij 0,bij,gij For PQ nodes those columns related to Q i and rows related to the voltage magnitude equation g i3 must be omitted. Blocking of node oriented variables e i ;f i ;Q i is important because this allows a fast and ecient block left or right preconditioning (see section 3). The linear system of equations of each Newton- Raphson step can be described in general form as follows: A 11 A 1 ::: A 1N A 1 A ::: A N Ax = b where A = A N1 A N ::: A NN (5) A is the above described Jacobian matrix, b is the mismatch vector, A ij are block submatrices of A determined with () and (4). This matrix A has the following properties: The block structure of A is identical with the structure of the complex nodal admittance matrix Y. It is very sparse, unsymmetric and indenite. Thus the CG method cannot be applied directly. 1 In order not to create confusion, a distinction between Newton-Raphson steps (Each step solves one approximation to the non-linear equations) and the term iteration to get the solution of one linear system with an iterative linear system solver is made throughout this paper. () (4) 3

3 3 BLOCK-DIAGONAL PRE- CONDITIONING OF LIN- EARIZED POWER FLOW EQUATIONS Using the Jacobian matrices as given in () and (4) in the non-stationary methods without any preconditioning leads to non-converging iterative behavior. In general preconditioning of a linear system of equations can be done as follows: (5) can be transformed with two matrices P L and P R as follows: P L A P R y = P L b x = P R y: (6) P L is called left and P R right preconditioning matrix. Preconditioning has the goal of making the new conditioned matrix P L AP R I where I is the unity matrix. In general, the better the unity matrix is approximated the faster the solution of the conditioned system. In the application discussed in this paper preconditioning has to be very fast because it has to be applied to each new Jacobian matrix. Thus one obvious choice is a block-diagonal preconditioning matrix P = 64 A,1 11 A,1... A,1 NN 3 75 (7) whose inverted diagonal blocks are identical with those of the original matrix A. The eort to compute P is small as compared to the computation of other, more sophisticated preconditioners such as ILU0 (factorization without consideration of ll-in terms). Only matrices of size 3 3 or must be inverted. Also, the eort per iteration for the nonstationary methods is smaller as compared to other preconditioners which usually do not allow an explicit inversion of a submatrix of the original matrix A. In the simulation runs this block diagonal preconditioning has been applied explicitely before the actual iterations are started. P can be applied as left- or right-preconditioning matrix. Both cases have been simulated, see section 4... From a theoretical point of view the following can be observed: Using P as left preconditioning matrix leads to the following preconditioned linear system of equations: P Ax=P b (8) Using P as right preconditioning matrix leads to the following preconditioned linear system of equations: AP y =b (9) x =P y (8) and (9) can be written in more general form as follows: A 0 x 0 = b 0 (10) where A 0 = P A; b 0 = P b; x 0 = x for left preconditioning and A 0 = AP; b 0 = b; x 0 = y for right preconditioning. In both (8) and (9) A 0 has unity block diagonal matrices and numerically modied (as compared to A) block-o-diagonal matrices. This preconditioned matrix A 0 approximates the desired unity matrix which is the goal of preconditioning. For direct methods with exact numeric precision both (8) and (9) show the same result for x. Applying iterative methods, however, leads to very different iterative convergence processes for both conditioning methods. Today no practical theorem exists for large linear systems which predicts the best preconditioning method. Only the simulation can show which combination of preconditioning/iterative method is best. However, since a norm of the right hand side vector of the conditioned system, i.e. P b (8) for left conditioning and b for right-conditioning (9) is used to stop the iterative convergence process, dierently scaled mismatch vectors are used. This scaling has a consequence on the convergence criterion of the iterative method which is the norm: k~rk = kb0, A 0 x 0 k kb 0 k (11) In this paper only block-diagonal preconditioning has been used. 4 APPLICATION OF NON- STATIONARY ITERATIVE METHODS TO POWER FLOW PROBLEMS 4.1 Implementation aspects The non-stationary iterative methods have been implemented in Matlab 4.c. Direct methods have also been simulated to allow comparisons. Matlab has the following implementation for a direct solution of a linear system with an unsymmetric matrix A 0 : with A 0 x = b 0 ~L U = ~ A 3

4 where ~ L is a permuted left matrix and U is an upper matrix as produced by the Matlab-\lu" algorithm. ~ A is the matrix A 0, permuted columnwise similar to the Tinney scheme. Tests were run on a Sparc 10. The power ow (PF) equations have been programmed according to (1)... (4). The networks used in the simulation runs have the following characteristics: No. No. Dimension Non-zeros Busses Branches Jacobian Jacobian Both 33 and 685 networks data are based on real power system network parts in Europe and in the U.S.A. Other, smaller networks (7, 57 bus networks) have also been used. However, simulation results have shown that no generalized conclusions can be drawn from these small networks. The following parameters have been used for all power ow simulations: Flat start for voltages (1 p.u) and reactive generator power (0); Max. number NR steps = 5; PF convergence tolerance = 0.01 p.u. (1 ka, 0.01 kv ) (max. mismatch norm is taken); Max. number of iterations for each iterative solver per NR step = dimension of A; convergence tolerance of the normalized k~rk = mismatch max =100, i.e. the convergence tolerance of the iterative solver is dependent on the maximum mismatch of the previously computed NR step; In addition, for BICGSTAB only, if abs k~rk (k), k~rk (k,1) mismatch max 10,5 stop after k iterations. Start the iterative solution with x 0 = b 0 (since A 0 I). The following table summarizes all used abbreviations: Abbreviation Description LU Direct solution BICG Bi-Conjugate Gradients BICGSTAB Bi-Conjugate Gradients Stabilized CGS Conjugate Gradients Squared QMR Quasi Minimum Residual GMRES(m) Generalized Minimum Residual (restarted after m iterations) CG Conjugate Gradients NPC No preconditioning LPC Left preconditioning RPC Right preconditioning max. mismatch jbijmax max PC-G max. conditioned mismatch jb 0 i j max k~rk NR Newton-Raphson cpu CPU seconds iter Number of iterations per NR step NC Not Converged case Doing this can reduce the number of iterations signicantly. 4. Simulation results 4..1 Direct methods The following table indicates the computational effort for the direct solution of the \Current mismatch power ow". Each iteration includes setup of mismatches and Jacobian matrix, factorization and forward/backward substitution. 33/685-bus networks NR 33: 6.1 s 685: 16.5 s Step cpu cpu e Non-stationary iterative methods Fig. 1 shows the total CPU time of various nonstationary iterative methods for the 33 bus network. Cpu time [s] LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC Figure 1: CPU times for 33 bus network Figs. and 3 show the total CPU time and the total number of iterations (i.e the sum of all iterations over all NR steps) of the various non-stationary iterative methods for the 685 bus network. Some conclusions can be drawn from these test runs: BICGSTAB_RPC The direct method is faster than any of the iterative methods. Note, however, that only the direct method is performance-optimized within Matlab. In Fig. 3, one \iteration" in LU NPC corresponds to one NR step. Right (block-diagonal) preconditioning diverges with QMR and CGS. CGS_LPC CGS_RPC QMR_LPC 4

5 Cpu time [s] LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC BICGSTAB_RPC CGS_LPC CGS_RPC QMR_LPC GMRES(1)_LPC GMRES(1)_RPC NC Figure : CPU times for 685 bus network The most successful method, i.e. BICG- STAB RPC is analyzed in more detail in the following tables. NR 33 bus network: BICGSTAB RPC: 7.4 s Step iter cpu The 0 in the column \iter" indicates that the chosen initial solution point, i.e. x 0 = b 0 satises the chosen iterative convergence criterion and no iterations are necessary. 7.4 s is the total CPU time. No of iterations LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC BICGSTAB_RPC CGS_LPC Figure 3: Total no of iterations for 685 bus network The performance and robustness of the GM- RES(m) methods varies extremely with m and the type of network: For the 33 network, GM- RES(m) converges only for high values of m (right preconditioning: m 50, left preconditioning: m 150). However, the 685 network converges with m 10. The best run is achieved with m = 1, right preconditioning. BICGSTAB RPC is the fastest iterative method for all networks, see also Figs. 5 and 6. It shows consistently robust convergence, i.e. all cases converged and the number of NR steps was never larger than the direct method. Left (block-diagonal) preconditioning takes more iterations than BICGSTAB RPC and converges in all cases. Hybrid methods, i.e. combinations of the above mentioned methods are possible. However, the scope of this paper is to show a comparison of accepted and well dened non-stationary iterative methods. CGS_RPC NC QMR_LPC GMRES(1)_LPC GMRES(1)_RPC NR 685 bus network: BICGSTAB RPC: 38.6 s Step iter cpu To get more insight into the overall convergence of selected methods Figs show the norms, the max. mismatch and the conditioned max. mismatch during all iterations for all NR steps. The straight lines indicate the end/beginning of one NR step. Note that the max. mismatch (indicated with the dashed and dashed-dotted line in the gures) is usually not computed during the iterations. It is only displayed here to see the convergence behavior of the various iterative methods and the solution accuracy of the original non-linear equations during all iterations. Fig. 4 shows the s over all iterations for the 33 network, CGS LPC. It clearly shows very noisy behavior as compared to BICGSTAB RPC (see Fig. 5). However, it converges also to the desired tolerance. In Fig. 5 and 6, a dramatic visual convergence dierence can be observed between the 33 and the 685 network for the same method (BICGSTAB RPC). Fig. 7 BICG LPC shows noisy downward trend for the norm and the maximum mismatch. Fig. 8 shows BICGSTAB LPC. The problem converges using a large number of NR steps. Fig. 9 shows the convergence behavior for the 685 network, using GMRES(1) RPC. This method is the onlyone beside CG minimizing k~rk. Therefore the s decrease monotonically during the iterations of each NR step. Also, one can observe the typical restart behavior of the GMRES method: Every 1 iterations, a mismatch step can be seen. However, due to the large computational eort per iteration, this method is slower than BICGSTAB RPC. 5

6 Figure 4: CGS LPC - 33 bus network: Residual (70.8 s CPU time) Figure 7: BICG LPC bus network: and (6.6 s CPU time) 10 3 max PC G Figure 5: BICGSTAB RPC - 33 bus network: and (7.4 s CPU time) Figure 8: BICGSTAB LPC bus network:, max-pc-g and (5.3 s CPU time) Figure 6: BICGSTAB RPC bus network: and (38.6 s CPU time) Figure 9: GMRES(1) RPC bus network: and (64.7 s CPU time) 6

7 5 CONCLUSIONS References In this paper implementation aspects of nonstationary iterative methods have been presented. Emphasis has been given to a comparison of the direct and various iterative methods based on the \current mismatch power ow". Due to the fact, that the Jacobian derived from the non-linear power ow equations is unsymmetric and indenite CG cannot be applied. Preconditioning of matrices is key to successful convergence of all iterative methods applied to practical power system problems. In this paper a diagonal block preconditioning scheme has been applied. This scheme is fast and leads for many methods to successful convergence. Simulation results applied to networks up to size 685 busses allow the following conclusions: The BICGSTAB RPC \Bi-Conjugate Gradients Stabilized right preconditioned" algorithm is the most robust algorithm for the solution of the \Current mismatch power ow" equations with a Newton- Raphson approach. The CPU times for a one CPU implementation of this algorithm are times slower than those of direct solution methods, all implemented in Matlab 4.c. The \Conjugate Gradient Squared" algorithm together with left block diagonal preconditioning (CGS LPC) shows good robustness, however, it is slower than BICGSTAB RPC. The \Generalized minimum " algorithm together with right block diagonal preconditioning (GMRES(m) RPC) shows smooth convergence properties even for quite low values of m. Although being slower than BICGSTAB RPC and CGS LPC this method is very appealing, because it minimizes the s from one iteration to the next in the same way as the \Conjugate Gradients" (CG) method does for symmetric and positive denite matrices. The algorithms presented in this paper together with the diagonal block preconditioning show almost perfect parallelism and can be implemented easily in a parallel CPU environment. A parallel implementation will reduce the total computation time significantly. This paper presents for the rst time practical applications of these non-stationary methods to large size non-linear power ow problems with unsymmetric and indenite Jacobian matrices. The implementation results in this paper show the relative performance and robustness of the various non-stationary methods applied to the power ow problem. It can be foreseen that the use of more sophisticated preconditioners, a deeper understanding of the characteristics of these methods applied to the power ow and the use of parallel CPU environments will further improve performance and robustness. [1] W.F. Tinney and J.W. Walker, Direct solutions of sparse network equations by optimally ordered triangular factorization, Proceedings of the IEEE,Vol. 55, Nov. 1967, pp [] F. D. Galiana, H. Javidi., S. McFee, On the application of a preconditioned conjugate gradient algorithm to power network analysis, IEEE Transactions on Power Systems, Vol. 9, No., May 1994, pp [3] H. Mori, H. Tanaka, A preconditioned fast decoupled power ow method for contingency screening, IEEE Power Industry Computer Applications Conference, Salt Lake City, May 7-1, 1995, pp [4] F. Alvarado, D. Hasan, S. Harmohan, Application of conjugate gradient method to power system least squares problems, SIAM conference on Linear Algebra, Snowbird, Colorado, June 1994 [5] F. Alvarado, H. Da~g and M. ten Bruggencate, Block-Bordered Diagonalization and Parallel Iterative Solvers, Colorado Conference on Iterative Methods, Breckenridge, Colorado, April 5{ 9, [6] A. Semlyen, Fundamental concepts of a Krylov subspace power ow methodology IEEE SM PWRS, IEEE/PES Summer Meeting, July 3-7, 1995, Portland, OR [7] R. Barrett, M. Berry, T. Chang, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, Ch. Romine, H. Van der Vorst, Templates for the solution of linear systems: Building blocks for iterative methods, SIAM, Philadelphia, Pennsylvania, 1993 (ftp netlib.cs.utk.edu; cd templates; get templates.ps) [8] J. R. Shewchuk, An introduction to the conjugate gradient method without the agonizing pain, School of Computer Science, Carnegie Mellon University, Pittsburgh, Ed. 1 1, August 1994 (ftp 4 warp.cs.cmu.edu; cd quake-papers; get painlessconjugate-gradient.ps) [9] M. R. Hestenes, E. Stiefel, Methods of conjugate gradients for solving linear systems, J. Res. National bureau of standards, Vol. 49, p , 195 [10] R. Bacher, Computer aided power ow software engineering and code generation, IEEE Power Industry Computer Applications Conference, Salt Lake City, May 7-1, 1995, pp

Nonsymmetric Problems. Abstract. The eect of a threshold variant TPABLO of the permutation

Nonsymmetric Problems. Abstract. The eect of a threshold variant TPABLO of the permutation Threshold Ordering for Preconditioning Nonsymmetric Problems Michele Benzi 1, Hwajeong Choi 2, Daniel B. Szyld 2? 1 CERFACS, 42 Ave. G. Coriolis, 31057 Toulouse Cedex, France (benzi@cerfacs.fr) 2 Department

More information

Techniques for Optimizing FEM/MoM Codes

Techniques for Optimizing FEM/MoM Codes Techniques for Optimizing FEM/MoM Codes Y. Ji, T. H. Hubing, and H. Wang Electromagnetic Compatibility Laboratory Department of Electrical & Computer Engineering University of Missouri-Rolla Rolla, MO

More information

THE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECH- NOLOGY ITERATIVE LINEAR SOLVERS

THE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECH- NOLOGY ITERATIVE LINEAR SOLVERS ITERATIVE LIEAR SOLVERS. Objectives The goals of the laboratory workshop are as follows: to learn basic properties of iterative methods for solving linear least squares problems, to study the properties

More information

Cots Sparse Matrix Utilization in Distribution Power Flow Applications

Cots Sparse Matrix Utilization in Distribution Power Flow Applications Cots Sparse Matrix Utilization in Distribution Power Flow Applications Dino Ablakovic, Izudin Dzafic and Hans-Theo Neisius Abstract Sparse matrix is used in many modern industrial applications, to provide

More information

Xinyu Dou Acoustics Technology Center, Motorola, Inc., Schaumburg, Illinois 60196

Xinyu Dou Acoustics Technology Center, Motorola, Inc., Schaumburg, Illinois 60196 A unified boundary element method for the analysis of sound and shell-like structure interactions. II. Efficient solution techniques Shaohai Chen and Yijun Liu a) Department of Mechanical Engineering,

More information

Sparse Linear Systems

Sparse Linear Systems 1 Sparse Linear Systems Rob H. Bisseling Mathematical Institute, Utrecht University Course Introduction Scientific Computing February 22, 2018 2 Outline Iterative solution methods 3 A perfect bipartite

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 20: Sparse Linear Systems; Direct Methods vs. Iterative Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 26

More information

UNSTRUCTURED GRIDS ON NURBS SURFACES. The surface grid can be generated either in a parameter. surfaces. Generating grids in a parameter space is

UNSTRUCTURED GRIDS ON NURBS SURFACES. The surface grid can be generated either in a parameter. surfaces. Generating grids in a parameter space is UNSTRUCTURED GRIDS ON NURBS SURFACES Jamshid Samareh-Abolhassani 1 Abstract A simple and ecient computational method is presented for unstructured surface grid generation. This method is built upon an

More information

Analysis of the GCR method with mixed precision arithmetic using QuPAT

Analysis of the GCR method with mixed precision arithmetic using QuPAT Analysis of the GCR method with mixed precision arithmetic using QuPAT Tsubasa Saito a,, Emiko Ishiwata b, Hidehiko Hasegawa c a Graduate School of Science, Tokyo University of Science, 1-3 Kagurazaka,

More information

NAG Library Function Document nag_sparse_sym_sol (f11jec)

NAG Library Function Document nag_sparse_sym_sol (f11jec) f11 Large Scale Linear Systems f11jec NAG Library Function Document nag_sparse_sym_sol (f11jec) 1 Purpose nag_sparse_sym_sol (f11jec) solves a real sparse symmetric system of linear equations, represented

More information

Parallel Numerical Algorithms

Parallel Numerical Algorithms Parallel Numerical Algorithms Chapter 4 Sparse Linear Systems Section 4.3 Iterative Methods Michael T. Heath and Edgar Solomonik Department of Computer Science University of Illinois at Urbana-Champaign

More information

(Sparse) Linear Solvers

(Sparse) Linear Solvers (Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert

More information

Iterative Solver Benchmark Jack Dongarra, Victor Eijkhout, Henk van der Vorst 2001/01/14 1 Introduction The traditional performance measurement for co

Iterative Solver Benchmark Jack Dongarra, Victor Eijkhout, Henk van der Vorst 2001/01/14 1 Introduction The traditional performance measurement for co Iterative Solver Benchmark Jack Dongarra, Victor Eijkhout, Henk van der Vorst 2001/01/14 1 Introduction The traditional performance measurement for computers on scientic application has been the Linpack

More information

Implicit schemes for wave models

Implicit schemes for wave models Implicit schemes for wave models Mathieu Dutour Sikirić Rudjer Bo sković Institute, Croatia and Universität Rostock April 17, 2013 I. Wave models Stochastic wave modelling Oceanic models are using grids

More information

Re-Dispatching Generation to Increase Power System Security Margin and Support Low Voltage Bus

Re-Dispatching Generation to Increase Power System Security Margin and Support Low Voltage Bus 496 IEEE TRANSACTIONS ON POWER SYSTEMS, VOL 15, NO 2, MAY 2000 Re-Dispatching Generation to Increase Power System Security Margin and Support Low Voltage Bus Ronghai Wang, Student Member, IEEE, and Robert

More information

Improvements of the Discrete Dipole Approximation method

Improvements of the Discrete Dipole Approximation method arxiv:physics/0006064v1 [physics.ao-ph] 26 Jun 2000 Improvements of the Discrete Dipole Approximation method Piotr J. Flatau Scripps Institution of Oceanography, University of California, San Diego, La

More information

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND Student Submission for the 5 th OpenFOAM User Conference 2017, Wiesbaden - Germany: SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND TESSA UROIĆ Faculty of Mechanical Engineering and Naval Architecture, Ivana

More information

Comparison of Two Stationary Iterative Methods

Comparison of Two Stationary Iterative Methods Comparison of Two Stationary Iterative Methods Oleksandra Osadcha Faculty of Applied Mathematics Silesian University of Technology Gliwice, Poland Email:olekosa@studentpolslpl Abstract This paper illustrates

More information

nag sparse nsym sol (f11dec)

nag sparse nsym sol (f11dec) f11 Sparse Linear Algebra f11dec nag sparse nsym sol (f11dec) 1. Purpose nag sparse nsym sol (f11dec) solves a real sparse nonsymmetric system of linear equations, represented in coordinate storage format,

More information

GPU-based Parallel Reservoir Simulators

GPU-based Parallel Reservoir Simulators GPU-based Parallel Reservoir Simulators Zhangxin Chen 1, Hui Liu 1, Song Yu 1, Ben Hsieh 1 and Lei Shao 1 Key words: GPU computing, reservoir simulation, linear solver, parallel 1 Introduction Nowadays

More information

NAG Fortran Library Routine Document F11DSF.1

NAG Fortran Library Routine Document F11DSF.1 NAG Fortran Library Routine Document Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

Developing object-oriented parallel iterative methods

Developing object-oriented parallel iterative methods Int. J. High Performance Computing and Networking, Vol. 1, Nos. 1/2/3, 2004 85 Developing object-oriented parallel iterative methods Chakib Ouarraui and David Kaeli* Department of Electrical and Computer

More information

Sparse Matrix Libraries in C++ for High Performance. Architectures. ferent sparse matrix data formats in order to best

Sparse Matrix Libraries in C++ for High Performance. Architectures. ferent sparse matrix data formats in order to best Sparse Matrix Libraries in C++ for High Performance Architectures Jack Dongarra xz, Andrew Lumsdaine, Xinhui Niu Roldan Pozo z, Karin Remington x x Oak Ridge National Laboratory z University oftennessee

More information

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear

More information

Report of Linear Solver Implementation on GPU

Report of Linear Solver Implementation on GPU Report of Linear Solver Implementation on GPU XIANG LI Abstract As the development of technology and the linear equation solver is used in many aspects such as smart grid, aviation and chemical engineering,

More information

arxiv: v1 [cs.ms] 2 Jun 2016

arxiv: v1 [cs.ms] 2 Jun 2016 Parallel Triangular Solvers on GPU Zhangxin Chen, Hui Liu, and Bo Yang University of Calgary 2500 University Dr NW, Calgary, AB, Canada, T2N 1N4 {zhachen,hui.j.liu,yang6}@ucalgary.ca arxiv:1606.00541v1

More information

GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013

GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013 GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS Kyle Spagnoli Research Engineer @ EM Photonics 3/20/2013 INTRODUCTION» Sparse systems» Iterative solvers» High level benchmarks»

More information

Algebraic Iterative Methods for Computed Tomography

Algebraic Iterative Methods for Computed Tomography Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic

More information

Tools and Libraries for Parallel Sparse Matrix Computations. Edmond Chow and Yousef Saad. University of Minnesota. Minneapolis, MN

Tools and Libraries for Parallel Sparse Matrix Computations. Edmond Chow and Yousef Saad. University of Minnesota. Minneapolis, MN Tools and Libraries for Parallel Sparse Matrix Computations Edmond Chow and Yousef Saad Department of Computer Science, and Minnesota Supercomputer Institute University of Minnesota Minneapolis, MN 55455

More information

Efficient Linear System Solvers for Mesh Processing

Efficient Linear System Solvers for Mesh Processing Efficient Linear System Solvers for Mesh Processing Mario Botsch, David Bommes, and Leif Kobbelt Computer Graphics Group, RWTH Aachen Technical University Abstract. The use of polygonal mesh representations

More information

Solution of 2D Euler Equations and Application to Airfoil Design

Solution of 2D Euler Equations and Application to Airfoil Design WDS'6 Proceedings of Contributed Papers, Part I, 47 52, 26. ISBN 8-86732-84-3 MATFYZPRESS Solution of 2D Euler Equations and Application to Airfoil Design J. Šimák Charles University, Faculty of Mathematics

More information

(Sparse) Linear Solvers

(Sparse) Linear Solvers (Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 1 Don t you just invert

More information

Iterative Sparse Triangular Solves for Preconditioning

Iterative Sparse Triangular Solves for Preconditioning Euro-Par 2015, Vienna Aug 24-28, 2015 Iterative Sparse Triangular Solves for Preconditioning Hartwig Anzt, Edmond Chow and Jack Dongarra Incomplete Factorization Preconditioning Incomplete LU factorizations

More information

NAG Library Function Document nag_sparse_nsym_sol (f11dec)

NAG Library Function Document nag_sparse_nsym_sol (f11dec) f11 Large Scale Linear Systems NAG Library Function Document nag_sparse_nsym_sol () 1 Purpose nag_sparse_nsym_sol () solves a real sparse nonsymmetric system of linear equations, represented in coordinate

More information

PARDISO Version Reference Sheet Fortran

PARDISO Version Reference Sheet Fortran PARDISO Version 5.0.0 1 Reference Sheet Fortran CALL PARDISO(PT, MAXFCT, MNUM, MTYPE, PHASE, N, A, IA, JA, 1 PERM, NRHS, IPARM, MSGLVL, B, X, ERROR, DPARM) 1 Please note that this version differs significantly

More information

Solving Sparse Linear Systems. Forward and backward substitution for solving lower or upper triangular systems

Solving Sparse Linear Systems. Forward and backward substitution for solving lower or upper triangular systems AMSC 6 /CMSC 76 Advanced Linear Numerical Analysis Fall 7 Direct Solution of Sparse Linear Systems and Eigenproblems Dianne P. O Leary c 7 Solving Sparse Linear Systems Assumed background: Gauss elimination

More information

1 2 (3 + x 3) x 2 = 1 3 (3 + x 1 2x 3 ) 1. 3 ( 1 x 2) (3 + x(0) 3 ) = 1 2 (3 + 0) = 3. 2 (3 + x(0) 1 2x (0) ( ) = 1 ( 1 x(0) 2 ) = 1 3 ) = 1 3

1 2 (3 + x 3) x 2 = 1 3 (3 + x 1 2x 3 ) 1. 3 ( 1 x 2) (3 + x(0) 3 ) = 1 2 (3 + 0) = 3. 2 (3 + x(0) 1 2x (0) ( ) = 1 ( 1 x(0) 2 ) = 1 3 ) = 1 3 6 Iterative Solvers Lab Objective: Many real-world problems of the form Ax = b have tens of thousands of parameters Solving such systems with Gaussian elimination or matrix factorizations could require

More information

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanfordedu) February 6, 2018 Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 In the

More information

paper, we focussed on the GMRES(m) which is improved GMRES(Generalized Minimal RESidual method), and developed its library on distributed memory machi

paper, we focussed on the GMRES(m) which is improved GMRES(Generalized Minimal RESidual method), and developed its library on distributed memory machi Performance of Automatically Tuned Parallel GMRES(m) Method on Distributed Memory Machines Hisayasu KURODA 1?, Takahiro KATAGIRI 12, and Yasumasa KANADA 3 1 Department of Information Science, Graduate

More information

Preliminary Investigations on Resilient Parallel Numerical Linear Algebra Solvers

Preliminary Investigations on Resilient Parallel Numerical Linear Algebra Solvers SIAM EX14 Workshop July 7, Chicago - IL reliminary Investigations on Resilient arallel Numerical Linear Algebra Solvers HieACS Inria roject Joint Inria-CERFACS lab INRIA Bordeaux Sud-Ouest Luc Giraud joint

More information

A Compiler for Parallel Finite Element Methods. with Domain-Decomposed Unstructured Meshes JONATHAN RICHARD SHEWCHUK AND OMAR GHATTAS

A Compiler for Parallel Finite Element Methods. with Domain-Decomposed Unstructured Meshes JONATHAN RICHARD SHEWCHUK AND OMAR GHATTAS Contemporary Mathematics Volume 00, 0000 A Compiler for Parallel Finite Element Methods with Domain-Decomposed Unstructured Meshes JONATHAN RICHARD SHEWCHUK AND OMAR GHATTAS December 11, 1993 Abstract.

More information

A Simple and Direct Approach for Unbalanced Radial Distribution System three phase Load Flow Solution

A Simple and Direct Approach for Unbalanced Radial Distribution System three phase Load Flow Solution Research Journal of Applied Sciences, Engineering and Technology 2(5): 452-459, 2010 ISSN: 2040-7467 Maxwell Scientific Organization, 2010 Submitted Date: May 16, 2010 Accepted Date: May 27, 2010 Published

More information

Contents. I The Basic Framework for Stationary Problems 1

Contents. I The Basic Framework for Stationary Problems 1 page v Preface xiii I The Basic Framework for Stationary Problems 1 1 Some model PDEs 3 1.1 Laplace s equation; elliptic BVPs... 3 1.1.1 Physical experiments modeled by Laplace s equation... 5 1.2 Other

More information

HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE

HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER AVISHA DHISLE PRERIT RODNEY ADHISLE PRODNEY 15618: PARALLEL COMPUTER ARCHITECTURE PROF. BRYANT PROF. KAYVON LET S

More information

Parallel solution for finite element linear systems of. equations on workstation cluster *

Parallel solution for finite element linear systems of. equations on workstation cluster * Aug. 2009, Volume 6, No.8 (Serial No.57) Journal of Communication and Computer, ISSN 1548-7709, USA Parallel solution for finite element linear systems of equations on workstation cluster * FU Chao-jiang

More information

Edge detection based on single layer CNN simulator using RK6(4)

Edge detection based on single layer CNN simulator using RK6(4) Edge detection based on single layer CNN simulator using RK64) Osama H. Abdelwahed 1, and M. El-Sayed Wahed 1 Mathematics Department, Faculty of Science, Suez Canal University, Egypt Department of Computer

More information

Efficient Use of Iterative Solvers in Nested Topology Optimization

Efficient Use of Iterative Solvers in Nested Topology Optimization Efficient Use of Iterative Solvers in Nested Topology Optimization Oded Amir, Mathias Stolpe and Ole Sigmund Technical University of Denmark Department of Mathematics Department of Mechanical Engineering

More information

Practical aspects of running EH3D E. Haber Λ June 4, 2001 Abstract This document discuss of practical aspects of running the program EH3D. The goal of

Practical aspects of running EH3D E. Haber Λ June 4, 2001 Abstract This document discuss of practical aspects of running the program EH3D. The goal of Practical aspects of running EH3D E. Haber Λ June 4, Abstract This document discuss of practical aspects of running the program EH3D. The goal of this document is to help practitioners determine and tune

More information

Solving IK problems for open chains using optimization methods

Solving IK problems for open chains using optimization methods Proceedings of the International Multiconference on Computer Science and Information Technology pp. 933 937 ISBN 978-83-60810-14-9 ISSN 1896-7094 Solving IK problems for open chains using optimization

More information

A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation

A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation Amir Nejat * and Carl Ollivier-Gooch Department of Mechanical Engineering, The University of British Columbia, BC V6T 1Z4, Canada

More information

Comparison of parallel preconditioners for a Newton-Krylov flow solver

Comparison of parallel preconditioners for a Newton-Krylov flow solver Comparison of parallel preconditioners for a Newton-Krylov flow solver Jason E. Hicken, Michal Osusky, and David W. Zingg 1Introduction Analysis of the results from the AIAA Drag Prediction workshops (Mavriplis

More information

A Novel Method for Power-Flow Solution of Radial Distribution Networks

A Novel Method for Power-Flow Solution of Radial Distribution Networks A Novel Method for Power-Flow Solution of Radial Distribution Networks 1 Narinder Singh, 2 Prof. Rajni Bala 1 Student-M.Tech(Power System), 2 Professor(Power System) BBSBEC, Fatehgarh Sahib, Punjab Abstract

More information

CS 542G: Solving Sparse Linear Systems

CS 542G: Solving Sparse Linear Systems CS 542G: Solving Sparse Linear Systems Robert Bridson November 26, 2008 1 Direct Methods We have already derived several methods for solving a linear system, say Ax = b, or the related leastsquares problem

More information

Lecture 9. Introduction to Numerical Techniques

Lecture 9. Introduction to Numerical Techniques Lecture 9. Introduction to Numerical Techniques Ivan Papusha CDS270 2: Mathematical Methods in Control and System Engineering May 27, 2015 1 / 25 Logistics hw8 (last one) due today. do an easy problem

More information

Performance Evaluation of a New Parallel Preconditioner

Performance Evaluation of a New Parallel Preconditioner Performance Evaluation of a New Parallel Preconditioner Keith D. Gremban Gary L. Miller Marco Zagha School of Computer Science Carnegie Mellon University 5 Forbes Avenue Pittsburgh PA 15213 Abstract The

More information

Uppsala University Department of Information technology. Hands-on 1: Ill-conditioning = x 2

Uppsala University Department of Information technology. Hands-on 1: Ill-conditioning = x 2 Uppsala University Department of Information technology Hands-on : Ill-conditioning Exercise (Ill-conditioned linear systems) Definition A system of linear equations is said to be ill-conditioned when

More information

RAPID COMPUTATION OF THE DISCRETE FOURIER TRANSFORM*

RAPID COMPUTATION OF THE DISCRETE FOURIER TRANSFORM* SIAM J. ScI. COMPUT. Vol. 17, No. 4, pp. 913-919, July 1996 1996 Society for Industrial and Applied Mathematics O08 RAPID COMPUTATION OF THE DISCRETE FOURIER TRANSFORM* CHRIS ANDERSON AND MARIE DILLON

More information

Parallel resolution of sparse linear systems by mixing direct and iterative methods

Parallel resolution of sparse linear systems by mixing direct and iterative methods Parallel resolution of sparse linear systems by mixing direct and iterative methods Phyleas Meeting, Bordeaux J. Gaidamour, P. Hénon, J. Roman, Y. Saad LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix

More information

Parallel ILU Ordering and Convergence Relationships: Numerical Experiments

Parallel ILU Ordering and Convergence Relationships: Numerical Experiments NASA/CR-00-2119 ICASE Report No. 00-24 Parallel ILU Ordering and Convergence Relationships: Numerical Experiments David Hysom and Alex Pothen Old Dominion University, Norfolk, Virginia Institute for Computer

More information

Aim. Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity. Structure and matrix sparsity: Overview

Aim. Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity. Structure and matrix sparsity: Overview Aim Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk What should a 2-hour PhD lecture on structure

More information

A General Sparse Sparse Linear System Solver and Its Application in OpenFOAM

A General Sparse Sparse Linear System Solver and Its Application in OpenFOAM Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe A General Sparse Sparse Linear System Solver and Its Application in OpenFOAM Murat Manguoglu * Middle East Technical University,

More information

The Solution of Systems of Linear Equations. using the Conjugate Gradient Method. Jean-Guy Schneider, Edgar F.A. Lederer, Peter Schwab.

The Solution of Systems of Linear Equations. using the Conjugate Gradient Method. Jean-Guy Schneider, Edgar F.A. Lederer, Peter Schwab. The Solution of Systems of Linear Equations using the Conjugate Gradient Method on the Parallel MUSIC-System Jean-Guy Schneider, Edgar F.A. Lederer, Peter Schwab Abstract The solution of large sparse systems

More information

Highly Parallel Multigrid Solvers for Multicore and Manycore Processors

Highly Parallel Multigrid Solvers for Multicore and Manycore Processors Highly Parallel Multigrid Solvers for Multicore and Manycore Processors Oleg Bessonov (B) Institute for Problems in Mechanics of the Russian Academy of Sciences, 101, Vernadsky Avenue, 119526 Moscow, Russia

More information

Lecture 15: More Iterative Ideas

Lecture 15: More Iterative Ideas Lecture 15: More Iterative Ideas David Bindel 15 Mar 2010 Logistics HW 2 due! Some notes on HW 2. Where we are / where we re going More iterative ideas. Intro to HW 3. More HW 2 notes See solution code!

More information

Lecture 27: Fast Laplacian Solvers

Lecture 27: Fast Laplacian Solvers Lecture 27: Fast Laplacian Solvers Scribed by Eric Lee, Eston Schweickart, Chengrun Yang November 21, 2017 1 How Fast Laplacian Solvers Work We want to solve Lx = b with L being a Laplacian matrix. Recall

More information

to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics

to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics An Application of Lagrangian Relaxation to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics M lardalen University SE-721

More information

All use is subject to licence, see For any commercial application, a separate licence must be signed.

All use is subject to licence, see   For any commercial application, a separate licence must be signed. HS PAKAGE SPEIFIATION HS 2007 1 SUMMARY This routine uses the Generalized Minimal Residual method with restarts every m iterations, GMRES(m), to solve the n n unsymmetric linear system Ax = b, optionally

More information

Accelerating a Simulation of Type I X ray Bursts from Accreting Neutron Stars Mark Mackey Professor Alexander Heger

Accelerating a Simulation of Type I X ray Bursts from Accreting Neutron Stars Mark Mackey Professor Alexander Heger Accelerating a Simulation of Type I X ray Bursts from Accreting Neutron Stars Mark Mackey Professor Alexander Heger The goal of my project was to develop an optimized linear system solver to shorten the

More information

EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI

EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI 1 Akshay N. Panajwar, 2 Prof.M.A.Shah Department of Computer Science and Engineering, Walchand College of Engineering,

More information

However, m pq is just an approximation of M pq. As it was pointed out by Lin [2], more precise approximation can be obtained by exact integration of t

However, m pq is just an approximation of M pq. As it was pointed out by Lin [2], more precise approximation can be obtained by exact integration of t FAST CALCULATION OF GEOMETRIC MOMENTS OF BINARY IMAGES Jan Flusser Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodarenskou vez 4, 82 08 Prague 8, Czech

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

On Improving Computational Efficiency of Simplified Fluid Flow Models

On Improving Computational Efficiency of Simplified Fluid Flow Models 1447 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 70, 2018 Guest Editors: Timothy G. Walmsley, Petar S. Varbanov, Rongxin Su, Jiří J. Klemeš Copyright 2018, AIDIC Servizi S.r.l. ISBN 978-88-95608-67-9;

More information

Implementation of QR Up- and Downdating on a. Massively Parallel Computer. Hans Bruun Nielsen z Mustafa Pnar z. July 8, 1996.

Implementation of QR Up- and Downdating on a. Massively Parallel Computer. Hans Bruun Nielsen z Mustafa Pnar z. July 8, 1996. Implementation of QR Up- and Downdating on a Massively Parallel Computer Claus Btsen y Per Christian Hansen y Kaj Madsen z Hans Bruun Nielsen z Mustafa Pnar z July 8, 1996 Abstract We describe an implementation

More information

ECEN 615 Methods of Electric Power Systems Analysis Lecture 13: Sparse Matrix Ordering, Sparse Vector Methods

ECEN 615 Methods of Electric Power Systems Analysis Lecture 13: Sparse Matrix Ordering, Sparse Vector Methods ECEN 615 Methods of Electric Power Systems Analysis Lecture 13: Sparse Matrix Ordering, Sparse Vector Methods Prof. Tom Overbye Dept. of Electrical and Computer Engineering Texas A&M University overbye@tamu.edu

More information

Lecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice

Lecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice Stat60/CS94: Randomized Algorithms for Matrices and Data Lecture 11-10/09/013 Lecture 11: Randomized Least-squares Approximation in Practice Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these

More information

Parallel Incomplete-LU and Cholesky Factorization in the Preconditioned Iterative Methods on the GPU

Parallel Incomplete-LU and Cholesky Factorization in the Preconditioned Iterative Methods on the GPU Parallel Incomplete-LU and Cholesky Factorization in the Preconditioned Iterative Methods on the GPU Maxim Naumov NVIDIA, 2701 San Tomas Expressway, Santa Clara, CA 95050 Abstract A novel algorithm for

More information

A parallel direct/iterative solver based on a Schur complement approach

A parallel direct/iterative solver based on a Schur complement approach A parallel direct/iterative solver based on a Schur complement approach Gene around the world at CERFACS Jérémie Gaidamour LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix project) February 29th, 2008

More information

Iterative Methods for Solving Linear Problems

Iterative Methods for Solving Linear Problems Iterative Methods for Solving Linear Problems When problems become too large (too many data points, too many model parameters), SVD and related approaches become impractical. Iterative Methods for Solving

More information

Image Registration with Automatic Computation of Gradients

Image Registration with Automatic Computation of Gradients Image Registration with Automatic Computation of Gradients Release 1.0 E. G. Kahn 1 and L. H. Staib 2 July 29, 2008 1 The Johns Hopkins University Applied Physics Laboratory, Laurel, Maryland 2 Yale University,

More information

Systems of Linear Equations and their Graphical Solution

Systems of Linear Equations and their Graphical Solution Proceedings of the World Congress on Engineering and Computer Science Vol I WCECS, - October,, San Francisco, USA Systems of Linear Equations and their Graphical Solution ISBN: 98-988-95-- ISSN: 8-958

More information

4.4 Example of [ Z Bus ] matrix formulation in the presence of mutual impedances

4.4 Example of [ Z Bus ] matrix formulation in the presence of mutual impedances 4.4 Example of [ Z matrix formulation in the presence of mutual impedances Consider the network shown in Fig. 4.32. Figure 4.32: The power system for [ Z example A tree for the network is shown in Fig.

More information

Journal of Engineering Research and Studies E-ISSN

Journal of Engineering Research and Studies E-ISSN Journal of Engineering Research and Studies E-ISS 0976-79 Research Article SPECTRAL SOLUTIO OF STEADY STATE CODUCTIO I ARBITRARY QUADRILATERAL DOMAIS Alavani Chitra R 1*, Joshi Pallavi A 1, S Pavitran

More information

Figure 6.1: Truss topology optimization diagram.

Figure 6.1: Truss topology optimization diagram. 6 Implementation 6.1 Outline This chapter shows the implementation details to optimize the truss, obtained in the ground structure approach, according to the formulation presented in previous chapters.

More information

Algebraic Iterative Methods for Computed Tomography

Algebraic Iterative Methods for Computed Tomography Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic

More information

Bias-Variance Tradeos Analysis Using Uniform CR Bound. Mohammad Usman, Alfred O. Hero, Jerey A. Fessler and W. L. Rogers. University of Michigan

Bias-Variance Tradeos Analysis Using Uniform CR Bound. Mohammad Usman, Alfred O. Hero, Jerey A. Fessler and W. L. Rogers. University of Michigan Bias-Variance Tradeos Analysis Using Uniform CR Bound Mohammad Usman, Alfred O. Hero, Jerey A. Fessler and W. L. Rogers University of Michigan ABSTRACT We quantify fundamental bias-variance tradeos for

More information

Sparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach

Sparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach 1 Sparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach David Greiner, Gustavo Montero, Gabriel Winter Institute of Intelligent Systems and Numerical Applications in Engineering (IUSIANI)

More information

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations Ben Jeuris and Raf Vandebril KU Leuven, Dept. of Computer Science, 3001 Leuven(Heverlee), Belgium {ben.jeuris,raf.vandebril}@cs.kuleuven.be

More information

NEW ADVANCES IN GPU LINEAR ALGEBRA

NEW ADVANCES IN GPU LINEAR ALGEBRA GTC 2012: NEW ADVANCES IN GPU LINEAR ALGEBRA Kyle Spagnoli EM Photonics 5/16/2012 QUICK ABOUT US» HPC/GPU Consulting Firm» Specializations in:» Electromagnetics» Image Processing» Fluid Dynamics» Linear

More information

Developing a High Performance Software Library with MPI and CUDA for Matrix Computations

Developing a High Performance Software Library with MPI and CUDA for Matrix Computations Developing a High Performance Software Library with MPI and CUDA for Matrix Computations Bogdan Oancea 1, Tudorel Andrei 2 1 Nicolae Titulescu University of Bucharest, e-mail: bogdanoancea@univnt.ro, Calea

More information

Performance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development

Performance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Performance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development M. Serdar Celebi

More information

General purpose fast decoupled power flow

General purpose fast decoupled power flow General purpose fast decoupled power flow Prof. J. Nanda P.R. Bijwe J. Henry V. Bapi Raju Indexing terms: Power system protection, Modelling Abstract: A general purpose fast decoupled power flow model

More information

A hybrid GMRES and TV-norm based method for image restoration

A hybrid GMRES and TV-norm based method for image restoration A hybrid GMRES and TV-norm based method for image restoration D. Calvetti a, B. Lewis b and L. Reichel c a Department of Mathematics, Case Western Reserve University, Cleveland, OH 44106 b Rocketcalc,

More information

Wei Shu and Min-You Wu. Abstract. partitioning patterns, and communication optimization to achieve a speedup.

Wei Shu and Min-You Wu. Abstract. partitioning patterns, and communication optimization to achieve a speedup. Sparse Implementation of Revised Simplex Algorithms on Parallel Computers Wei Shu and Min-You Wu Abstract Parallelizing sparse simplex algorithms is one of the most challenging problems. Because of very

More information

AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS

AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS Seyed Abolfazl Shahzadehfazeli 1, Zainab Haji Abootorabi,3 1 Parallel Processing Laboratory, Yazd University,

More information

Blocked Schur Algorithms for Computing the Matrix Square Root. Deadman, Edvin and Higham, Nicholas J. and Ralha, Rui. MIMS EPrint: 2012.

Blocked Schur Algorithms for Computing the Matrix Square Root. Deadman, Edvin and Higham, Nicholas J. and Ralha, Rui. MIMS EPrint: 2012. Blocked Schur Algorithms for Computing the Matrix Square Root Deadman, Edvin and Higham, Nicholas J. and Ralha, Rui 2013 MIMS EPrint: 2012.26 Manchester Institute for Mathematical Sciences School of Mathematics

More information

HIPS : a parallel hybrid direct/iterative solver based on a Schur complement approach

HIPS : a parallel hybrid direct/iterative solver based on a Schur complement approach HIPS : a parallel hybrid direct/iterative solver based on a Schur complement approach Mini-workshop PHyLeaS associated team J. Gaidamour, P. Hénon July 9, 28 HIPS : an hybrid direct/iterative solver /

More information

User Manual for the Complex Conjugate Gradient Methods Library CCGPAK 2.0

User Manual for the Complex Conjugate Gradient Methods Library CCGPAK 2.0 User Manual for the Complex Conjugate Gradient Methods Library CCGPAK 2.0 Piotr J. Flatau Scripps Institution of Oceanography University of California San Diego La Jolla, CA92093 email: pflatau@ucsd.edu

More information

Computational issues in linear programming

Computational issues in linear programming Computational issues in linear programming Julian Hall School of Mathematics University of Edinburgh 15th May 2007 Computational issues in linear programming Overview Introduction to linear programming

More information

Load Flow Calculation For Electrical Power System Based On Run Length Encoding Algorithm

Load Flow Calculation For Electrical Power System Based On Run Length Encoding Algorithm International Conference on Renewable Energies and Power Quality (ICREPQ 4) Cordoba (Spain), 8 th to 0 th April, 204 exçxãtuäx XÇxÜzç tçw céãxü dâtä àç ]ÉâÜÇtÄ (RE&PQJ) ISSN 272-038 X,.2, April 204 Load

More information

On Testing a Linear State Estimator Using Hardware in the Loop Testing Facilities

On Testing a Linear State Estimator Using Hardware in the Loop Testing Facilities 1 On Testing a Linear State Estimator Using Hardware in the Loop Testing Facilities Vladimir Terzija, Pawel Regulski, Alexandru Nechifor The University of Manchester, UK Presenter: Junbo Zhao Contents

More information