Cpu time [s] BICGSTAB_RPC CGS_RPC BICGSTAB_LPC BICG_RPC BICG_LPC LU_NPC
|
|
- Bruno Walsh
- 6 years ago
- Views:
Transcription
1 Application of Non-stationary Iterative Methods to an Exact Newton-Raphson Solution Process for Power Flow Equations Rainer Bacher, Eric Bullinger Swiss Federal Institute of Technology (ETH), CH-809 Zurich, Switzerland ABSTRACT The AC power ow is usually solved by the Newton- Raphson solution method. The main step is the linearization of the non-linear power ow equations and the subsequent solution of this linear system. The characteristics of this linear system of equations vary for dierent power ow implementations: Symmetric/unsymmetric, positive denite/indenite system matrices can result. Based on these characteristics dierent direct and iterative linear system solvers can be used to maximize performance and solution robustness. In this paper results are given of nonstationary iterative methods for unsymmetric, indefinite linear systems derived from power ow equations. Keywords Load ow analysis, Modelling and simulation, Non-stationary iterative methods 1 INTRODUCTION The power ow is a very well known algorithmic problem which is usually solved by the Newton- Raphson solution method. The linearization of the non-linear power ow equations yields a linear system to be solved by an appropriate linear system solver. All Newton-Raphson (NR) based power ow algorithms have in common that one large or two smaller size linear systems of equations must be solved during each Newton-Raphson iterative step. Beside the well known direct solution of a linear system of equations [1], the solution based on socalled non-stationary iterative methods has recently appeared in power applications. [] describes the rst application of the Conjugate Gradients (CG) method to the decoupled power ow. In [3] another application of CG methods applied to a static security power ow problem is described. For the fast decoupled power ow both papers state a signicant performance improvement of CG based methods compared to a direct solution. From all power ow approaches known today only the decoupled power ow satises the CG conditions of positive denite and symmetric linear system matrices without applying equation transformations. Practical usage of the conjugate gradient methods is only reached when preconditioning is applied to the linear system of equations. Good preconditioning will group the eigenvalues of the transformed linear system matrix together and will thus result in faster convergence. Again, there is a natural t between the preconditioned CG method and the decoupled power ow, since the decoupled power has constant linear system matrices. Thus, for every power ow this preconditioning matrix must be computed only once and remains constant for all Newton-Raphson steps. [4], [5] emphasize the fact that only good preconditioners allow an ecient implementation of CG methods to power ow equations. Several preconditioners such as ILU(0) (no ll-in factorization) or ILU(1) (approximate factorization with 1 neighbor ll-in term) together with block-bordered matrix permutation have been applied successfully to fast decoupled power ows with large networks. Recently, in [6], the rst application of nonstationary, iterative methods to the non-linear power ow problem has been described. A derivation of the \Krylov subspace power ow methodology" applied to the power ow problem is given to introduce power system application developers to the mathematical problem. The main distinction to the CG methods lies in the fact that the \Krylov subspace power ow methodology" is also applicable to unsymmetric, indenite linear system matrices. The so-called KSPF (Krylov Subspace Power Flow) which is derived in [6] does not need any explicit computation of the Jacobian terms during the iterations and power ow steps. Good convergence is shown with networks up to size 57 busses. In this paper implementation aspects of nonstationary methods with larger networks are discussed which are in principle applicable to all power ow formulations with unsymmetric and indenite linear system matrices. Mathematicians have de- 1
2 rived several methods to solve this type of linear system of equations: QMR (Quasi Minimal Residual), GMRES (Generalized Minimal Residual), BICG (Bi- Conjugate Gradients), CGS(Conjugate Gradients Squared), BICGSTAB (Bi-Conjugate Gradients Stabilized) are distinctly dierent methods for the solution of this class of problems. A good summary of these methods can be found in [7]. The GMRES-algorithm is the generalization of the CG algorithm for unsymmetric and indenite linear system matrices. Both algorithms have in common that the solution error (, see section 3) decreases from one iteration to the next and the exact solution is obtained within a given maximum number of iterations. This assumes exact numeric precision. All other non-stationary iterative methods use combinations of CG-like concepts and heuristics to obtain a solution for the linear system. As a consequence the of these methods is not guaranteed to decrease during the iterations. This paper gives details related to a successful computer implementation of non-stationary iterative methods applied to the non-linear power ow equations (1) and simulation results which should allow to get insight into the strengths and weaknesses of non-stationary iterative methods. After a problem denition in section, a special preconditioning of the linear system related to the Newton-Raphson solution process is given in section 3. In section 4, simulation results are discussed for various non-stationary methods applied to networks with 33 and 685 busses. The paper ends with conclusions in section 5 and the references. The direct solution based on factorization and a subsequent forward/backward substitution is not discussed in this paper, see [1]. Also, the mathematical theory for the non-stationary iterative method is not given in this paper due to space reasons, see [7]. An excellent reference to CG is [8]. The original description of CG is [9]. PROBLEM STATEMENT This paper deals with the iterative behavior of the solution of the linear system within the Newton- Raphson solution process for the non-linear power ow equations. The power ow equations for each node i are as follows: g i1 g i = e ip i +f i Q i e +f i i =,e iq i +f i P i e +f i i, P N j=1 (g ije j, b ij f j ) = 0 P N, j=1 (g ijf j + b ij e j ) = 0 i = 0 g i3 =,e i, f i + V (1) e i ;f i represent the real and imaginary part of nodal voltages, P i ;Q i represent the active and reactive nodal power, g ij ;b ij the real and imaginary part of the nodal admittance matrix Y ij, N = total number of nodes. For PQ-nodes the third equation g i3 is not needed. This power ow equations set (1) is called the \Current mismatch power ow" [10] since mainly nodal current equations are formulated. Note that the reactive powers Q i are unknown variables at PV generator nodes. These variables must be updated during all Newton-Raphson steps and linear system iterations 1. The Jacobian matrix terms are as follows: Jacobian diagonal block A ii 1 3 x = ei x = fi x = Qi ii, gii ii, bii ii + bii,ii, gii f i e i +f i, e i e i +f i,ei,fi 0 ii =, Pi (e i,f)+e i if i Q i (e i +f i ) ii =, Qi (f i,e i )+e if i P i (e i +f i ) (3) O-diagonal block A ij ;i 1 3 x=ej(i6=j) x=fj(i6=j) x=qj(i6=j),gij bij 0,bij,gij For PQ nodes those columns related to Q i and rows related to the voltage magnitude equation g i3 must be omitted. Blocking of node oriented variables e i ;f i ;Q i is important because this allows a fast and ecient block left or right preconditioning (see section 3). The linear system of equations of each Newton- Raphson step can be described in general form as follows: A 11 A 1 ::: A 1N A 1 A ::: A N Ax = b where A = A N1 A N ::: A NN (5) A is the above described Jacobian matrix, b is the mismatch vector, A ij are block submatrices of A determined with () and (4). This matrix A has the following properties: The block structure of A is identical with the structure of the complex nodal admittance matrix Y. It is very sparse, unsymmetric and indenite. Thus the CG method cannot be applied directly. 1 In order not to create confusion, a distinction between Newton-Raphson steps (Each step solves one approximation to the non-linear equations) and the term iteration to get the solution of one linear system with an iterative linear system solver is made throughout this paper. () (4) 3
3 3 BLOCK-DIAGONAL PRE- CONDITIONING OF LIN- EARIZED POWER FLOW EQUATIONS Using the Jacobian matrices as given in () and (4) in the non-stationary methods without any preconditioning leads to non-converging iterative behavior. In general preconditioning of a linear system of equations can be done as follows: (5) can be transformed with two matrices P L and P R as follows: P L A P R y = P L b x = P R y: (6) P L is called left and P R right preconditioning matrix. Preconditioning has the goal of making the new conditioned matrix P L AP R I where I is the unity matrix. In general, the better the unity matrix is approximated the faster the solution of the conditioned system. In the application discussed in this paper preconditioning has to be very fast because it has to be applied to each new Jacobian matrix. Thus one obvious choice is a block-diagonal preconditioning matrix P = 64 A,1 11 A,1... A,1 NN 3 75 (7) whose inverted diagonal blocks are identical with those of the original matrix A. The eort to compute P is small as compared to the computation of other, more sophisticated preconditioners such as ILU0 (factorization without consideration of ll-in terms). Only matrices of size 3 3 or must be inverted. Also, the eort per iteration for the nonstationary methods is smaller as compared to other preconditioners which usually do not allow an explicit inversion of a submatrix of the original matrix A. In the simulation runs this block diagonal preconditioning has been applied explicitely before the actual iterations are started. P can be applied as left- or right-preconditioning matrix. Both cases have been simulated, see section 4... From a theoretical point of view the following can be observed: Using P as left preconditioning matrix leads to the following preconditioned linear system of equations: P Ax=P b (8) Using P as right preconditioning matrix leads to the following preconditioned linear system of equations: AP y =b (9) x =P y (8) and (9) can be written in more general form as follows: A 0 x 0 = b 0 (10) where A 0 = P A; b 0 = P b; x 0 = x for left preconditioning and A 0 = AP; b 0 = b; x 0 = y for right preconditioning. In both (8) and (9) A 0 has unity block diagonal matrices and numerically modied (as compared to A) block-o-diagonal matrices. This preconditioned matrix A 0 approximates the desired unity matrix which is the goal of preconditioning. For direct methods with exact numeric precision both (8) and (9) show the same result for x. Applying iterative methods, however, leads to very different iterative convergence processes for both conditioning methods. Today no practical theorem exists for large linear systems which predicts the best preconditioning method. Only the simulation can show which combination of preconditioning/iterative method is best. However, since a norm of the right hand side vector of the conditioned system, i.e. P b (8) for left conditioning and b for right-conditioning (9) is used to stop the iterative convergence process, dierently scaled mismatch vectors are used. This scaling has a consequence on the convergence criterion of the iterative method which is the norm: k~rk = kb0, A 0 x 0 k kb 0 k (11) In this paper only block-diagonal preconditioning has been used. 4 APPLICATION OF NON- STATIONARY ITERATIVE METHODS TO POWER FLOW PROBLEMS 4.1 Implementation aspects The non-stationary iterative methods have been implemented in Matlab 4.c. Direct methods have also been simulated to allow comparisons. Matlab has the following implementation for a direct solution of a linear system with an unsymmetric matrix A 0 : with A 0 x = b 0 ~L U = ~ A 3
4 where ~ L is a permuted left matrix and U is an upper matrix as produced by the Matlab-\lu" algorithm. ~ A is the matrix A 0, permuted columnwise similar to the Tinney scheme. Tests were run on a Sparc 10. The power ow (PF) equations have been programmed according to (1)... (4). The networks used in the simulation runs have the following characteristics: No. No. Dimension Non-zeros Busses Branches Jacobian Jacobian Both 33 and 685 networks data are based on real power system network parts in Europe and in the U.S.A. Other, smaller networks (7, 57 bus networks) have also been used. However, simulation results have shown that no generalized conclusions can be drawn from these small networks. The following parameters have been used for all power ow simulations: Flat start for voltages (1 p.u) and reactive generator power (0); Max. number NR steps = 5; PF convergence tolerance = 0.01 p.u. (1 ka, 0.01 kv ) (max. mismatch norm is taken); Max. number of iterations for each iterative solver per NR step = dimension of A; convergence tolerance of the normalized k~rk = mismatch max =100, i.e. the convergence tolerance of the iterative solver is dependent on the maximum mismatch of the previously computed NR step; In addition, for BICGSTAB only, if abs k~rk (k), k~rk (k,1) mismatch max 10,5 stop after k iterations. Start the iterative solution with x 0 = b 0 (since A 0 I). The following table summarizes all used abbreviations: Abbreviation Description LU Direct solution BICG Bi-Conjugate Gradients BICGSTAB Bi-Conjugate Gradients Stabilized CGS Conjugate Gradients Squared QMR Quasi Minimum Residual GMRES(m) Generalized Minimum Residual (restarted after m iterations) CG Conjugate Gradients NPC No preconditioning LPC Left preconditioning RPC Right preconditioning max. mismatch jbijmax max PC-G max. conditioned mismatch jb 0 i j max k~rk NR Newton-Raphson cpu CPU seconds iter Number of iterations per NR step NC Not Converged case Doing this can reduce the number of iterations signicantly. 4. Simulation results 4..1 Direct methods The following table indicates the computational effort for the direct solution of the \Current mismatch power ow". Each iteration includes setup of mismatches and Jacobian matrix, factorization and forward/backward substitution. 33/685-bus networks NR 33: 6.1 s 685: 16.5 s Step cpu cpu e Non-stationary iterative methods Fig. 1 shows the total CPU time of various nonstationary iterative methods for the 33 bus network. Cpu time [s] LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC Figure 1: CPU times for 33 bus network Figs. and 3 show the total CPU time and the total number of iterations (i.e the sum of all iterations over all NR steps) of the various non-stationary iterative methods for the 685 bus network. Some conclusions can be drawn from these test runs: BICGSTAB_RPC The direct method is faster than any of the iterative methods. Note, however, that only the direct method is performance-optimized within Matlab. In Fig. 3, one \iteration" in LU NPC corresponds to one NR step. Right (block-diagonal) preconditioning diverges with QMR and CGS. CGS_LPC CGS_RPC QMR_LPC 4
5 Cpu time [s] LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC BICGSTAB_RPC CGS_LPC CGS_RPC QMR_LPC GMRES(1)_LPC GMRES(1)_RPC NC Figure : CPU times for 685 bus network The most successful method, i.e. BICG- STAB RPC is analyzed in more detail in the following tables. NR 33 bus network: BICGSTAB RPC: 7.4 s Step iter cpu The 0 in the column \iter" indicates that the chosen initial solution point, i.e. x 0 = b 0 satises the chosen iterative convergence criterion and no iterations are necessary. 7.4 s is the total CPU time. No of iterations LU_NPC BICG_LPC BICG_RPC BICGSTAB_LPC BICGSTAB_RPC CGS_LPC Figure 3: Total no of iterations for 685 bus network The performance and robustness of the GM- RES(m) methods varies extremely with m and the type of network: For the 33 network, GM- RES(m) converges only for high values of m (right preconditioning: m 50, left preconditioning: m 150). However, the 685 network converges with m 10. The best run is achieved with m = 1, right preconditioning. BICGSTAB RPC is the fastest iterative method for all networks, see also Figs. 5 and 6. It shows consistently robust convergence, i.e. all cases converged and the number of NR steps was never larger than the direct method. Left (block-diagonal) preconditioning takes more iterations than BICGSTAB RPC and converges in all cases. Hybrid methods, i.e. combinations of the above mentioned methods are possible. However, the scope of this paper is to show a comparison of accepted and well dened non-stationary iterative methods. CGS_RPC NC QMR_LPC GMRES(1)_LPC GMRES(1)_RPC NR 685 bus network: BICGSTAB RPC: 38.6 s Step iter cpu To get more insight into the overall convergence of selected methods Figs show the norms, the max. mismatch and the conditioned max. mismatch during all iterations for all NR steps. The straight lines indicate the end/beginning of one NR step. Note that the max. mismatch (indicated with the dashed and dashed-dotted line in the gures) is usually not computed during the iterations. It is only displayed here to see the convergence behavior of the various iterative methods and the solution accuracy of the original non-linear equations during all iterations. Fig. 4 shows the s over all iterations for the 33 network, CGS LPC. It clearly shows very noisy behavior as compared to BICGSTAB RPC (see Fig. 5). However, it converges also to the desired tolerance. In Fig. 5 and 6, a dramatic visual convergence dierence can be observed between the 33 and the 685 network for the same method (BICGSTAB RPC). Fig. 7 BICG LPC shows noisy downward trend for the norm and the maximum mismatch. Fig. 8 shows BICGSTAB LPC. The problem converges using a large number of NR steps. Fig. 9 shows the convergence behavior for the 685 network, using GMRES(1) RPC. This method is the onlyone beside CG minimizing k~rk. Therefore the s decrease monotonically during the iterations of each NR step. Also, one can observe the typical restart behavior of the GMRES method: Every 1 iterations, a mismatch step can be seen. However, due to the large computational eort per iteration, this method is slower than BICGSTAB RPC. 5
6 Figure 4: CGS LPC - 33 bus network: Residual (70.8 s CPU time) Figure 7: BICG LPC bus network: and (6.6 s CPU time) 10 3 max PC G Figure 5: BICGSTAB RPC - 33 bus network: and (7.4 s CPU time) Figure 8: BICGSTAB LPC bus network:, max-pc-g and (5.3 s CPU time) Figure 6: BICGSTAB RPC bus network: and (38.6 s CPU time) Figure 9: GMRES(1) RPC bus network: and (64.7 s CPU time) 6
7 5 CONCLUSIONS References In this paper implementation aspects of nonstationary iterative methods have been presented. Emphasis has been given to a comparison of the direct and various iterative methods based on the \current mismatch power ow". Due to the fact, that the Jacobian derived from the non-linear power ow equations is unsymmetric and indenite CG cannot be applied. Preconditioning of matrices is key to successful convergence of all iterative methods applied to practical power system problems. In this paper a diagonal block preconditioning scheme has been applied. This scheme is fast and leads for many methods to successful convergence. Simulation results applied to networks up to size 685 busses allow the following conclusions: The BICGSTAB RPC \Bi-Conjugate Gradients Stabilized right preconditioned" algorithm is the most robust algorithm for the solution of the \Current mismatch power ow" equations with a Newton- Raphson approach. The CPU times for a one CPU implementation of this algorithm are times slower than those of direct solution methods, all implemented in Matlab 4.c. The \Conjugate Gradient Squared" algorithm together with left block diagonal preconditioning (CGS LPC) shows good robustness, however, it is slower than BICGSTAB RPC. The \Generalized minimum " algorithm together with right block diagonal preconditioning (GMRES(m) RPC) shows smooth convergence properties even for quite low values of m. Although being slower than BICGSTAB RPC and CGS LPC this method is very appealing, because it minimizes the s from one iteration to the next in the same way as the \Conjugate Gradients" (CG) method does for symmetric and positive denite matrices. The algorithms presented in this paper together with the diagonal block preconditioning show almost perfect parallelism and can be implemented easily in a parallel CPU environment. A parallel implementation will reduce the total computation time significantly. This paper presents for the rst time practical applications of these non-stationary methods to large size non-linear power ow problems with unsymmetric and indenite Jacobian matrices. The implementation results in this paper show the relative performance and robustness of the various non-stationary methods applied to the power ow problem. It can be foreseen that the use of more sophisticated preconditioners, a deeper understanding of the characteristics of these methods applied to the power ow and the use of parallel CPU environments will further improve performance and robustness. [1] W.F. Tinney and J.W. Walker, Direct solutions of sparse network equations by optimally ordered triangular factorization, Proceedings of the IEEE,Vol. 55, Nov. 1967, pp [] F. D. Galiana, H. Javidi., S. McFee, On the application of a preconditioned conjugate gradient algorithm to power network analysis, IEEE Transactions on Power Systems, Vol. 9, No., May 1994, pp [3] H. Mori, H. Tanaka, A preconditioned fast decoupled power ow method for contingency screening, IEEE Power Industry Computer Applications Conference, Salt Lake City, May 7-1, 1995, pp [4] F. Alvarado, D. Hasan, S. Harmohan, Application of conjugate gradient method to power system least squares problems, SIAM conference on Linear Algebra, Snowbird, Colorado, June 1994 [5] F. Alvarado, H. Da~g and M. ten Bruggencate, Block-Bordered Diagonalization and Parallel Iterative Solvers, Colorado Conference on Iterative Methods, Breckenridge, Colorado, April 5{ 9, [6] A. Semlyen, Fundamental concepts of a Krylov subspace power ow methodology IEEE SM PWRS, IEEE/PES Summer Meeting, July 3-7, 1995, Portland, OR [7] R. Barrett, M. Berry, T. Chang, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, Ch. Romine, H. Van der Vorst, Templates for the solution of linear systems: Building blocks for iterative methods, SIAM, Philadelphia, Pennsylvania, 1993 (ftp netlib.cs.utk.edu; cd templates; get templates.ps) [8] J. R. Shewchuk, An introduction to the conjugate gradient method without the agonizing pain, School of Computer Science, Carnegie Mellon University, Pittsburgh, Ed. 1 1, August 1994 (ftp 4 warp.cs.cmu.edu; cd quake-papers; get painlessconjugate-gradient.ps) [9] M. R. Hestenes, E. Stiefel, Methods of conjugate gradients for solving linear systems, J. Res. National bureau of standards, Vol. 49, p , 195 [10] R. Bacher, Computer aided power ow software engineering and code generation, IEEE Power Industry Computer Applications Conference, Salt Lake City, May 7-1, 1995, pp
Nonsymmetric Problems. Abstract. The eect of a threshold variant TPABLO of the permutation
Threshold Ordering for Preconditioning Nonsymmetric Problems Michele Benzi 1, Hwajeong Choi 2, Daniel B. Szyld 2? 1 CERFACS, 42 Ave. G. Coriolis, 31057 Toulouse Cedex, France (benzi@cerfacs.fr) 2 Department
More informationTechniques for Optimizing FEM/MoM Codes
Techniques for Optimizing FEM/MoM Codes Y. Ji, T. H. Hubing, and H. Wang Electromagnetic Compatibility Laboratory Department of Electrical & Computer Engineering University of Missouri-Rolla Rolla, MO
More informationTHE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECH- NOLOGY ITERATIVE LINEAR SOLVERS
ITERATIVE LIEAR SOLVERS. Objectives The goals of the laboratory workshop are as follows: to learn basic properties of iterative methods for solving linear least squares problems, to study the properties
More informationCots Sparse Matrix Utilization in Distribution Power Flow Applications
Cots Sparse Matrix Utilization in Distribution Power Flow Applications Dino Ablakovic, Izudin Dzafic and Hans-Theo Neisius Abstract Sparse matrix is used in many modern industrial applications, to provide
More informationXinyu Dou Acoustics Technology Center, Motorola, Inc., Schaumburg, Illinois 60196
A unified boundary element method for the analysis of sound and shell-like structure interactions. II. Efficient solution techniques Shaohai Chen and Yijun Liu a) Department of Mechanical Engineering,
More informationSparse Linear Systems
1 Sparse Linear Systems Rob H. Bisseling Mathematical Institute, Utrecht University Course Introduction Scientific Computing February 22, 2018 2 Outline Iterative solution methods 3 A perfect bipartite
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 20: Sparse Linear Systems; Direct Methods vs. Iterative Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 26
More informationUNSTRUCTURED GRIDS ON NURBS SURFACES. The surface grid can be generated either in a parameter. surfaces. Generating grids in a parameter space is
UNSTRUCTURED GRIDS ON NURBS SURFACES Jamshid Samareh-Abolhassani 1 Abstract A simple and ecient computational method is presented for unstructured surface grid generation. This method is built upon an
More informationAnalysis of the GCR method with mixed precision arithmetic using QuPAT
Analysis of the GCR method with mixed precision arithmetic using QuPAT Tsubasa Saito a,, Emiko Ishiwata b, Hidehiko Hasegawa c a Graduate School of Science, Tokyo University of Science, 1-3 Kagurazaka,
More informationNAG Library Function Document nag_sparse_sym_sol (f11jec)
f11 Large Scale Linear Systems f11jec NAG Library Function Document nag_sparse_sym_sol (f11jec) 1 Purpose nag_sparse_sym_sol (f11jec) solves a real sparse symmetric system of linear equations, represented
More informationParallel Numerical Algorithms
Parallel Numerical Algorithms Chapter 4 Sparse Linear Systems Section 4.3 Iterative Methods Michael T. Heath and Edgar Solomonik Department of Computer Science University of Illinois at Urbana-Champaign
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert
More informationIterative Solver Benchmark Jack Dongarra, Victor Eijkhout, Henk van der Vorst 2001/01/14 1 Introduction The traditional performance measurement for co
Iterative Solver Benchmark Jack Dongarra, Victor Eijkhout, Henk van der Vorst 2001/01/14 1 Introduction The traditional performance measurement for computers on scientic application has been the Linpack
More informationImplicit schemes for wave models
Implicit schemes for wave models Mathieu Dutour Sikirić Rudjer Bo sković Institute, Croatia and Universität Rostock April 17, 2013 I. Wave models Stochastic wave modelling Oceanic models are using grids
More informationRe-Dispatching Generation to Increase Power System Security Margin and Support Low Voltage Bus
496 IEEE TRANSACTIONS ON POWER SYSTEMS, VOL 15, NO 2, MAY 2000 Re-Dispatching Generation to Increase Power System Security Margin and Support Low Voltage Bus Ronghai Wang, Student Member, IEEE, and Robert
More informationImprovements of the Discrete Dipole Approximation method
arxiv:physics/0006064v1 [physics.ao-ph] 26 Jun 2000 Improvements of the Discrete Dipole Approximation method Piotr J. Flatau Scripps Institution of Oceanography, University of California, San Diego, La
More informationSELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND
Student Submission for the 5 th OpenFOAM User Conference 2017, Wiesbaden - Germany: SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND TESSA UROIĆ Faculty of Mechanical Engineering and Naval Architecture, Ivana
More informationComparison of Two Stationary Iterative Methods
Comparison of Two Stationary Iterative Methods Oleksandra Osadcha Faculty of Applied Mathematics Silesian University of Technology Gliwice, Poland Email:olekosa@studentpolslpl Abstract This paper illustrates
More informationnag sparse nsym sol (f11dec)
f11 Sparse Linear Algebra f11dec nag sparse nsym sol (f11dec) 1. Purpose nag sparse nsym sol (f11dec) solves a real sparse nonsymmetric system of linear equations, represented in coordinate storage format,
More informationGPU-based Parallel Reservoir Simulators
GPU-based Parallel Reservoir Simulators Zhangxin Chen 1, Hui Liu 1, Song Yu 1, Ben Hsieh 1 and Lei Shao 1 Key words: GPU computing, reservoir simulation, linear solver, parallel 1 Introduction Nowadays
More informationNAG Fortran Library Routine Document F11DSF.1
NAG Fortran Library Routine Document Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent
More informationDeveloping object-oriented parallel iterative methods
Int. J. High Performance Computing and Networking, Vol. 1, Nos. 1/2/3, 2004 85 Developing object-oriented parallel iterative methods Chakib Ouarraui and David Kaeli* Department of Electrical and Computer
More informationSparse Matrix Libraries in C++ for High Performance. Architectures. ferent sparse matrix data formats in order to best
Sparse Matrix Libraries in C++ for High Performance Architectures Jack Dongarra xz, Andrew Lumsdaine, Xinhui Niu Roldan Pozo z, Karin Remington x x Oak Ridge National Laboratory z University oftennessee
More informationIterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms
Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear
More informationReport of Linear Solver Implementation on GPU
Report of Linear Solver Implementation on GPU XIANG LI Abstract As the development of technology and the linear equation solver is used in many aspects such as smart grid, aviation and chemical engineering,
More informationarxiv: v1 [cs.ms] 2 Jun 2016
Parallel Triangular Solvers on GPU Zhangxin Chen, Hui Liu, and Bo Yang University of Calgary 2500 University Dr NW, Calgary, AB, Canada, T2N 1N4 {zhachen,hui.j.liu,yang6}@ucalgary.ca arxiv:1606.00541v1
More informationGTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013
GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS Kyle Spagnoli Research Engineer @ EM Photonics 3/20/2013 INTRODUCTION» Sparse systems» Iterative solvers» High level benchmarks»
More informationAlgebraic Iterative Methods for Computed Tomography
Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic
More informationTools and Libraries for Parallel Sparse Matrix Computations. Edmond Chow and Yousef Saad. University of Minnesota. Minneapolis, MN
Tools and Libraries for Parallel Sparse Matrix Computations Edmond Chow and Yousef Saad Department of Computer Science, and Minnesota Supercomputer Institute University of Minnesota Minneapolis, MN 55455
More informationEfficient Linear System Solvers for Mesh Processing
Efficient Linear System Solvers for Mesh Processing Mario Botsch, David Bommes, and Leif Kobbelt Computer Graphics Group, RWTH Aachen Technical University Abstract. The use of polygonal mesh representations
More informationSolution of 2D Euler Equations and Application to Airfoil Design
WDS'6 Proceedings of Contributed Papers, Part I, 47 52, 26. ISBN 8-86732-84-3 MATFYZPRESS Solution of 2D Euler Equations and Application to Airfoil Design J. Šimák Charles University, Faculty of Mathematics
More information(Sparse) Linear Solvers
(Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 1 Don t you just invert
More informationIterative Sparse Triangular Solves for Preconditioning
Euro-Par 2015, Vienna Aug 24-28, 2015 Iterative Sparse Triangular Solves for Preconditioning Hartwig Anzt, Edmond Chow and Jack Dongarra Incomplete Factorization Preconditioning Incomplete LU factorizations
More informationNAG Library Function Document nag_sparse_nsym_sol (f11dec)
f11 Large Scale Linear Systems NAG Library Function Document nag_sparse_nsym_sol () 1 Purpose nag_sparse_nsym_sol () solves a real sparse nonsymmetric system of linear equations, represented in coordinate
More informationPARDISO Version Reference Sheet Fortran
PARDISO Version 5.0.0 1 Reference Sheet Fortran CALL PARDISO(PT, MAXFCT, MNUM, MTYPE, PHASE, N, A, IA, JA, 1 PERM, NRHS, IPARM, MSGLVL, B, X, ERROR, DPARM) 1 Please note that this version differs significantly
More informationSolving Sparse Linear Systems. Forward and backward substitution for solving lower or upper triangular systems
AMSC 6 /CMSC 76 Advanced Linear Numerical Analysis Fall 7 Direct Solution of Sparse Linear Systems and Eigenproblems Dianne P. O Leary c 7 Solving Sparse Linear Systems Assumed background: Gauss elimination
More information1 2 (3 + x 3) x 2 = 1 3 (3 + x 1 2x 3 ) 1. 3 ( 1 x 2) (3 + x(0) 3 ) = 1 2 (3 + 0) = 3. 2 (3 + x(0) 1 2x (0) ( ) = 1 ( 1 x(0) 2 ) = 1 3 ) = 1 3
6 Iterative Solvers Lab Objective: Many real-world problems of the form Ax = b have tens of thousands of parameters Solving such systems with Gaussian elimination or matrix factorizations could require
More informationLecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1
CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanfordedu) February 6, 2018 Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 In the
More informationpaper, we focussed on the GMRES(m) which is improved GMRES(Generalized Minimal RESidual method), and developed its library on distributed memory machi
Performance of Automatically Tuned Parallel GMRES(m) Method on Distributed Memory Machines Hisayasu KURODA 1?, Takahiro KATAGIRI 12, and Yasumasa KANADA 3 1 Department of Information Science, Graduate
More informationPreliminary Investigations on Resilient Parallel Numerical Linear Algebra Solvers
SIAM EX14 Workshop July 7, Chicago - IL reliminary Investigations on Resilient arallel Numerical Linear Algebra Solvers HieACS Inria roject Joint Inria-CERFACS lab INRIA Bordeaux Sud-Ouest Luc Giraud joint
More informationA Compiler for Parallel Finite Element Methods. with Domain-Decomposed Unstructured Meshes JONATHAN RICHARD SHEWCHUK AND OMAR GHATTAS
Contemporary Mathematics Volume 00, 0000 A Compiler for Parallel Finite Element Methods with Domain-Decomposed Unstructured Meshes JONATHAN RICHARD SHEWCHUK AND OMAR GHATTAS December 11, 1993 Abstract.
More informationA Simple and Direct Approach for Unbalanced Radial Distribution System three phase Load Flow Solution
Research Journal of Applied Sciences, Engineering and Technology 2(5): 452-459, 2010 ISSN: 2040-7467 Maxwell Scientific Organization, 2010 Submitted Date: May 16, 2010 Accepted Date: May 27, 2010 Published
More informationContents. I The Basic Framework for Stationary Problems 1
page v Preface xiii I The Basic Framework for Stationary Problems 1 1 Some model PDEs 3 1.1 Laplace s equation; elliptic BVPs... 3 1.1.1 Physical experiments modeled by Laplace s equation... 5 1.2 Other
More informationHYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE
HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER AVISHA DHISLE PRERIT RODNEY ADHISLE PRODNEY 15618: PARALLEL COMPUTER ARCHITECTURE PROF. BRYANT PROF. KAYVON LET S
More informationParallel solution for finite element linear systems of. equations on workstation cluster *
Aug. 2009, Volume 6, No.8 (Serial No.57) Journal of Communication and Computer, ISSN 1548-7709, USA Parallel solution for finite element linear systems of equations on workstation cluster * FU Chao-jiang
More informationEdge detection based on single layer CNN simulator using RK6(4)
Edge detection based on single layer CNN simulator using RK64) Osama H. Abdelwahed 1, and M. El-Sayed Wahed 1 Mathematics Department, Faculty of Science, Suez Canal University, Egypt Department of Computer
More informationEfficient Use of Iterative Solvers in Nested Topology Optimization
Efficient Use of Iterative Solvers in Nested Topology Optimization Oded Amir, Mathias Stolpe and Ole Sigmund Technical University of Denmark Department of Mathematics Department of Mechanical Engineering
More informationPractical aspects of running EH3D E. Haber Λ June 4, 2001 Abstract This document discuss of practical aspects of running the program EH3D. The goal of
Practical aspects of running EH3D E. Haber Λ June 4, Abstract This document discuss of practical aspects of running the program EH3D. The goal of this document is to help practitioners determine and tune
More informationSolving IK problems for open chains using optimization methods
Proceedings of the International Multiconference on Computer Science and Information Technology pp. 933 937 ISBN 978-83-60810-14-9 ISSN 1896-7094 Solving IK problems for open chains using optimization
More informationA High-Order Accurate Unstructured GMRES Solver for Poisson s Equation
A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation Amir Nejat * and Carl Ollivier-Gooch Department of Mechanical Engineering, The University of British Columbia, BC V6T 1Z4, Canada
More informationComparison of parallel preconditioners for a Newton-Krylov flow solver
Comparison of parallel preconditioners for a Newton-Krylov flow solver Jason E. Hicken, Michal Osusky, and David W. Zingg 1Introduction Analysis of the results from the AIAA Drag Prediction workshops (Mavriplis
More informationA Novel Method for Power-Flow Solution of Radial Distribution Networks
A Novel Method for Power-Flow Solution of Radial Distribution Networks 1 Narinder Singh, 2 Prof. Rajni Bala 1 Student-M.Tech(Power System), 2 Professor(Power System) BBSBEC, Fatehgarh Sahib, Punjab Abstract
More informationCS 542G: Solving Sparse Linear Systems
CS 542G: Solving Sparse Linear Systems Robert Bridson November 26, 2008 1 Direct Methods We have already derived several methods for solving a linear system, say Ax = b, or the related leastsquares problem
More informationLecture 9. Introduction to Numerical Techniques
Lecture 9. Introduction to Numerical Techniques Ivan Papusha CDS270 2: Mathematical Methods in Control and System Engineering May 27, 2015 1 / 25 Logistics hw8 (last one) due today. do an easy problem
More informationPerformance Evaluation of a New Parallel Preconditioner
Performance Evaluation of a New Parallel Preconditioner Keith D. Gremban Gary L. Miller Marco Zagha School of Computer Science Carnegie Mellon University 5 Forbes Avenue Pittsburgh PA 15213 Abstract The
More informationUppsala University Department of Information technology. Hands-on 1: Ill-conditioning = x 2
Uppsala University Department of Information technology Hands-on : Ill-conditioning Exercise (Ill-conditioned linear systems) Definition A system of linear equations is said to be ill-conditioned when
More informationRAPID COMPUTATION OF THE DISCRETE FOURIER TRANSFORM*
SIAM J. ScI. COMPUT. Vol. 17, No. 4, pp. 913-919, July 1996 1996 Society for Industrial and Applied Mathematics O08 RAPID COMPUTATION OF THE DISCRETE FOURIER TRANSFORM* CHRIS ANDERSON AND MARIE DILLON
More informationParallel resolution of sparse linear systems by mixing direct and iterative methods
Parallel resolution of sparse linear systems by mixing direct and iterative methods Phyleas Meeting, Bordeaux J. Gaidamour, P. Hénon, J. Roman, Y. Saad LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix
More informationParallel ILU Ordering and Convergence Relationships: Numerical Experiments
NASA/CR-00-2119 ICASE Report No. 00-24 Parallel ILU Ordering and Convergence Relationships: Numerical Experiments David Hysom and Alex Pothen Old Dominion University, Norfolk, Virginia Institute for Computer
More informationAim. Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity. Structure and matrix sparsity: Overview
Aim Structure and matrix sparsity: Part 1 The simplex method: Exploiting sparsity Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk What should a 2-hour PhD lecture on structure
More informationA General Sparse Sparse Linear System Solver and Its Application in OpenFOAM
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe A General Sparse Sparse Linear System Solver and Its Application in OpenFOAM Murat Manguoglu * Middle East Technical University,
More informationThe Solution of Systems of Linear Equations. using the Conjugate Gradient Method. Jean-Guy Schneider, Edgar F.A. Lederer, Peter Schwab.
The Solution of Systems of Linear Equations using the Conjugate Gradient Method on the Parallel MUSIC-System Jean-Guy Schneider, Edgar F.A. Lederer, Peter Schwab Abstract The solution of large sparse systems
More informationHighly Parallel Multigrid Solvers for Multicore and Manycore Processors
Highly Parallel Multigrid Solvers for Multicore and Manycore Processors Oleg Bessonov (B) Institute for Problems in Mechanics of the Russian Academy of Sciences, 101, Vernadsky Avenue, 119526 Moscow, Russia
More informationLecture 15: More Iterative Ideas
Lecture 15: More Iterative Ideas David Bindel 15 Mar 2010 Logistics HW 2 due! Some notes on HW 2. Where we are / where we re going More iterative ideas. Intro to HW 3. More HW 2 notes See solution code!
More informationLecture 27: Fast Laplacian Solvers
Lecture 27: Fast Laplacian Solvers Scribed by Eric Lee, Eston Schweickart, Chengrun Yang November 21, 2017 1 How Fast Laplacian Solvers Work We want to solve Lx = b with L being a Laplacian matrix. Recall
More informationto the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics
An Application of Lagrangian Relaxation to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics M lardalen University SE-721
More informationAll use is subject to licence, see For any commercial application, a separate licence must be signed.
HS PAKAGE SPEIFIATION HS 2007 1 SUMMARY This routine uses the Generalized Minimal Residual method with restarts every m iterations, GMRES(m), to solve the n n unsymmetric linear system Ax = b, optionally
More informationAccelerating a Simulation of Type I X ray Bursts from Accreting Neutron Stars Mark Mackey Professor Alexander Heger
Accelerating a Simulation of Type I X ray Bursts from Accreting Neutron Stars Mark Mackey Professor Alexander Heger The goal of my project was to develop an optimized linear system solver to shorten the
More informationEFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI
EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI 1 Akshay N. Panajwar, 2 Prof.M.A.Shah Department of Computer Science and Engineering, Walchand College of Engineering,
More informationHowever, m pq is just an approximation of M pq. As it was pointed out by Lin [2], more precise approximation can be obtained by exact integration of t
FAST CALCULATION OF GEOMETRIC MOMENTS OF BINARY IMAGES Jan Flusser Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodarenskou vez 4, 82 08 Prague 8, Czech
More informationContents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.
page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5
More informationOn Improving Computational Efficiency of Simplified Fluid Flow Models
1447 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 70, 2018 Guest Editors: Timothy G. Walmsley, Petar S. Varbanov, Rongxin Su, Jiří J. Klemeš Copyright 2018, AIDIC Servizi S.r.l. ISBN 978-88-95608-67-9;
More informationImplementation of QR Up- and Downdating on a. Massively Parallel Computer. Hans Bruun Nielsen z Mustafa Pnar z. July 8, 1996.
Implementation of QR Up- and Downdating on a Massively Parallel Computer Claus Btsen y Per Christian Hansen y Kaj Madsen z Hans Bruun Nielsen z Mustafa Pnar z July 8, 1996 Abstract We describe an implementation
More informationECEN 615 Methods of Electric Power Systems Analysis Lecture 13: Sparse Matrix Ordering, Sparse Vector Methods
ECEN 615 Methods of Electric Power Systems Analysis Lecture 13: Sparse Matrix Ordering, Sparse Vector Methods Prof. Tom Overbye Dept. of Electrical and Computer Engineering Texas A&M University overbye@tamu.edu
More informationLecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice
Stat60/CS94: Randomized Algorithms for Matrices and Data Lecture 11-10/09/013 Lecture 11: Randomized Least-squares Approximation in Practice Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these
More informationParallel Incomplete-LU and Cholesky Factorization in the Preconditioned Iterative Methods on the GPU
Parallel Incomplete-LU and Cholesky Factorization in the Preconditioned Iterative Methods on the GPU Maxim Naumov NVIDIA, 2701 San Tomas Expressway, Santa Clara, CA 95050 Abstract A novel algorithm for
More informationA parallel direct/iterative solver based on a Schur complement approach
A parallel direct/iterative solver based on a Schur complement approach Gene around the world at CERFACS Jérémie Gaidamour LaBRI and INRIA Bordeaux - Sud-Ouest (ScAlApplix project) February 29th, 2008
More informationIterative Methods for Solving Linear Problems
Iterative Methods for Solving Linear Problems When problems become too large (too many data points, too many model parameters), SVD and related approaches become impractical. Iterative Methods for Solving
More informationImage Registration with Automatic Computation of Gradients
Image Registration with Automatic Computation of Gradients Release 1.0 E. G. Kahn 1 and L. H. Staib 2 July 29, 2008 1 The Johns Hopkins University Applied Physics Laboratory, Laurel, Maryland 2 Yale University,
More informationSystems of Linear Equations and their Graphical Solution
Proceedings of the World Congress on Engineering and Computer Science Vol I WCECS, - October,, San Francisco, USA Systems of Linear Equations and their Graphical Solution ISBN: 98-988-95-- ISSN: 8-958
More information4.4 Example of [ Z Bus ] matrix formulation in the presence of mutual impedances
4.4 Example of [ Z matrix formulation in the presence of mutual impedances Consider the network shown in Fig. 4.32. Figure 4.32: The power system for [ Z example A tree for the network is shown in Fig.
More informationJournal of Engineering Research and Studies E-ISSN
Journal of Engineering Research and Studies E-ISS 0976-79 Research Article SPECTRAL SOLUTIO OF STEADY STATE CODUCTIO I ARBITRARY QUADRILATERAL DOMAIS Alavani Chitra R 1*, Joshi Pallavi A 1, S Pavitran
More informationFigure 6.1: Truss topology optimization diagram.
6 Implementation 6.1 Outline This chapter shows the implementation details to optimize the truss, obtained in the ground structure approach, according to the formulation presented in previous chapters.
More informationAlgebraic Iterative Methods for Computed Tomography
Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic
More informationBias-Variance Tradeos Analysis Using Uniform CR Bound. Mohammad Usman, Alfred O. Hero, Jerey A. Fessler and W. L. Rogers. University of Michigan
Bias-Variance Tradeos Analysis Using Uniform CR Bound Mohammad Usman, Alfred O. Hero, Jerey A. Fessler and W. L. Rogers University of Michigan ABSTRACT We quantify fundamental bias-variance tradeos for
More informationSparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach
1 Sparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach David Greiner, Gustavo Montero, Gabriel Winter Institute of Intelligent Systems and Numerical Applications in Engineering (IUSIANI)
More informationGeometric Mean Algorithms Based on Harmonic and Arithmetic Iterations
Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations Ben Jeuris and Raf Vandebril KU Leuven, Dept. of Computer Science, 3001 Leuven(Heverlee), Belgium {ben.jeuris,raf.vandebril}@cs.kuleuven.be
More informationNEW ADVANCES IN GPU LINEAR ALGEBRA
GTC 2012: NEW ADVANCES IN GPU LINEAR ALGEBRA Kyle Spagnoli EM Photonics 5/16/2012 QUICK ABOUT US» HPC/GPU Consulting Firm» Specializations in:» Electromagnetics» Image Processing» Fluid Dynamics» Linear
More informationDeveloping a High Performance Software Library with MPI and CUDA for Matrix Computations
Developing a High Performance Software Library with MPI and CUDA for Matrix Computations Bogdan Oancea 1, Tudorel Andrei 2 1 Nicolae Titulescu University of Bucharest, e-mail: bogdanoancea@univnt.ro, Calea
More informationPerformance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Performance Analysis of BLAS Libraries in SuperLU_DIST for SuperLU_MCDT (Multi Core Distributed) Development M. Serdar Celebi
More informationGeneral purpose fast decoupled power flow
General purpose fast decoupled power flow Prof. J. Nanda P.R. Bijwe J. Henry V. Bapi Raju Indexing terms: Power system protection, Modelling Abstract: A general purpose fast decoupled power flow model
More informationA hybrid GMRES and TV-norm based method for image restoration
A hybrid GMRES and TV-norm based method for image restoration D. Calvetti a, B. Lewis b and L. Reichel c a Department of Mathematics, Case Western Reserve University, Cleveland, OH 44106 b Rocketcalc,
More informationWei Shu and Min-You Wu. Abstract. partitioning patterns, and communication optimization to achieve a speedup.
Sparse Implementation of Revised Simplex Algorithms on Parallel Computers Wei Shu and Min-You Wu Abstract Parallelizing sparse simplex algorithms is one of the most challenging problems. Because of very
More informationAN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS
AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS Seyed Abolfazl Shahzadehfazeli 1, Zainab Haji Abootorabi,3 1 Parallel Processing Laboratory, Yazd University,
More informationBlocked Schur Algorithms for Computing the Matrix Square Root. Deadman, Edvin and Higham, Nicholas J. and Ralha, Rui. MIMS EPrint: 2012.
Blocked Schur Algorithms for Computing the Matrix Square Root Deadman, Edvin and Higham, Nicholas J. and Ralha, Rui 2013 MIMS EPrint: 2012.26 Manchester Institute for Mathematical Sciences School of Mathematics
More informationHIPS : a parallel hybrid direct/iterative solver based on a Schur complement approach
HIPS : a parallel hybrid direct/iterative solver based on a Schur complement approach Mini-workshop PHyLeaS associated team J. Gaidamour, P. Hénon July 9, 28 HIPS : an hybrid direct/iterative solver /
More informationUser Manual for the Complex Conjugate Gradient Methods Library CCGPAK 2.0
User Manual for the Complex Conjugate Gradient Methods Library CCGPAK 2.0 Piotr J. Flatau Scripps Institution of Oceanography University of California San Diego La Jolla, CA92093 email: pflatau@ucsd.edu
More informationComputational issues in linear programming
Computational issues in linear programming Julian Hall School of Mathematics University of Edinburgh 15th May 2007 Computational issues in linear programming Overview Introduction to linear programming
More informationLoad Flow Calculation For Electrical Power System Based On Run Length Encoding Algorithm
International Conference on Renewable Energies and Power Quality (ICREPQ 4) Cordoba (Spain), 8 th to 0 th April, 204 exçxãtuäx XÇxÜzç tçw céãxü dâtä àç ]ÉâÜÇtÄ (RE&PQJ) ISSN 272-038 X,.2, April 204 Load
More informationOn Testing a Linear State Estimator Using Hardware in the Loop Testing Facilities
1 On Testing a Linear State Estimator Using Hardware in the Loop Testing Facilities Vladimir Terzija, Pawel Regulski, Alexandru Nechifor The University of Manchester, UK Presenter: Junbo Zhao Contents
More information