SUMMARY. solve the matrix system using iterative solvers. We use the MUMPS codes and distribute the computation over many different

Size: px
Start display at page:

Download "SUMMARY. solve the matrix system using iterative solvers. We use the MUMPS codes and distribute the computation over many different"

Transcription

1 Forward Modelling and Inversion of Multi-Source TEM Data D. W. Oldenburg 1, E. Haber 2, and R. Shekhtman 1 1 University of British Columbia, Department of Earth & Ocean Sciences 2 Emory University, Atlanta, GA SUMMARY We present a practical formulation for forward modeling and inverting time domain data arising from multiple transmitters. The underpinning of our procedure is the ability to factor the forward modeling matrix and then solve our system using direct methods. We first formulate Maxwell s equations in terms of the magnetic field, H to provide a symmetric forward modeling operator. The problem is discretized using a finite volume technique in space and a backward Euler in time. The MUMPS software package is used to carry out a Cholesky decomposition of the forward operator, with the work distributed over an array of processors. The forward modeling is then quickly carried out using the factored operator. The time savings are considerable and they make the simulations of large ground or airborne data sets feasible and greatly increase our abilities to solve the 3D electromagnetic inverse problem in a reasonable time. The ability to use direct methods also greatly enhances our ability to carry out survey design problems. INTRODUCTION In previous research (Haber, Oldenburg, & Shekhtman, 2006) we developed an inversion algorithm that allowed us to invert data from a single, or a very few, transmitters. Unfortunately the computational demands of that algorithm were too large to invert typical ground or airborne surveys acquired from many source locations. The principal difficulty is the time required to solve the forward problem. Simulating data that arise from multi-sources can be computationally onerous because each transmitter requires that Maxwell s equations be solved. Usually this is done with iterative (eg. CG-type) algorithms and hence the computation time increases linearly with the number of transmitter locations. However, for surveys with a large number of sources, significant increases in efficiency can be had if the forward modeling matrix is factored. Factorization involves large computations and significant memory requirements. However, once this is accomplished, solving the factored system with a different right hand side proceeds very quickly. The idea of decomposing the matrix system and solving many right hand sides for different sources is not new (Dey & Morrison, 1979), and small problems have been solved in this manner. However, the matrices for 3D TEM problems have generally been considered to be too large to contemplate this approach. Over the last decade however, advances in mathematics and computing science have resulted in factorization algorithms that can be implemented on large scale computing systems (Amestoy, Guermouche, LExcellent, & Pralet, 2006). The efficacy of this approach depends upon the time required to factor the matrix compared to the time required to solve the matrix system using iterative solvers. We use the MUMPS codes and distribute the computation over many different processors. MAXWELL S EQUATIONS IN THE TIME DOMAIN The forward model consists of Maxwell s equations in time where the permeability is fixed but electrical conductivity can be highly discontinuous. To be more specific, we write Maxwell s equations as E (i) + µ H(i) t H (i) σe (i) ε E(i) t i = 1,...,N = 0, (1a) = s (i) r (t) (1b) (1c) over a domain Ω [0,t f ], where E (i) and H (i) are the electric and magnetic fields that correspond to the source s (i) r, µ is the permeability, σ is the conductivity, ε is the permittivity and s r is a source. The equations are given with boundary and initial conditions: n H (i) = 0 (2a) H(0,x) = H (i) 0 (2b) E(0, x) = 0 (2c) although other boundary and initial conditions could be used. Solving the forward problem As previously discussed in (Haber & Ascher, 2001; Haber, Ascher, Aruliah, & Oldenburg, 2000) consistent discretization of Maxwell s equation leads to a near-singular system. This is due to the rich null space of the curl operator. In our previous work we have turned to the A φ formulation to stabilize the system. While this approach is highly effective for iterative methods, it is not as effective for direct methods due to two main difficulties. First, the resulting system is not symmetric thus one is required to use the LU factorization rather than the Cholesky factorization; this doubles the amount of memory used for the solution of the system. Second, the number of unknowns is larger and this adds even more memory to the factorization. Keeping this in mind, we now work directly with the fields. Using backward Euler discretization in time with step size δt we obtain E i+1 + µ H i+1 H i δt H i+1 σe i+1 ε E i+1 E i δt = 0, (3) = s ri+1 (4)

2 Eliminating the electric field E i+1 from Maxwell s equations we obtain an equation for the magnetic field H i+1 (σ + δt 1 ε) 1 H i+1 + µ δt H i+1 = rhs i+1 (5) The above system can be near-singular when the conductivity is small and the time steps are large. To stabilize it we recall that we also have µh = 0 and therefore we can rewrite the system as (σ + δt 1 ε) 1 H i+1 µ µ 2 (σ + δt 1 ε) 1 (µh i+1 ) (6a) (6b) + µ δt H i+1 = rhs i+1 (6c) The new symmetric term µ µ 2 (σδt 1 ε) 1 (µh i+1 ) stabilizes the system without changing the solution. Upon discretization is space, we obtain the linear system A(σ)H i+1 := (C(σ,δt) + M)H i+1 = rhs (7) where the matrix C(σ,δt) is a discretization of the differential operator in (7). The matrix is SPD and depends on the conductivity and the time step δt. The matrix M depends on µ and δt. We now make an important observation which motivates our approach. We note that by using the same time steps δt, the linear system (7) is identical for all times and all sources. Thus, assuming that we decompose the system into A(σ) = LL (8) this factorization can be used to solve all the linear systems. Matrix factorization is an expensive computational process. If the number of sources and/or time steps is small then we may not benefit from this work and using iterative methods for the solution of the system can be superior. However, when the same forward modeling matrix needs to be solved many times and sources, the decomposition will be greatly superior to iterative techniques. The benefits of this are exacerbated when one proceeds to the inverse problem since the same factorization can be used for the computation of the gradient as well as for the solution of the linear system which arises at each Gauss-Newton iteration. Software for Matrix Factorization In recent years there has been a growing effort to obtain scalable matrix factorizations on parallel machines. We use the package Multi-frontal Massively Parallel Solver (MUMPS) by the CERFACS group (Amestoy et al., 2006). MUMPS is a package for solving systems of linear equations of the form Ax = b, where the matrix A is sparse and can be either unsymmetric, symmetric positive definite, or general symmetric. MUMPS uses a multifrontal technique which is a direct method based on either the LU or the LDL factorization of the matrix. It exploits both parallelism arising from sparsity in the matrix A and from dense factorizations kernels. The main features of the MUMPS package include the solution of the transposed system, input of the matrix in assembled format (distributed or centralized) or elemental format, error analysis, iterative refinement, scaling of the original matrix, and return of a Schur complement matrix. Finally, MUMPS is available in various arithmetics (real or complex, single or double precision). The software is written in Fortran 90. The parallel version of MUMPS requires MPI for message passing and makes use of the BLAS, BLACS, and ScaLAPACK libraries. We have tested MUMPS on a cluster of PCs under Linux. MUMPS distributes the work tasks among the processors, but an identified processor (the host) is required to perform most of the analysis phase, distribute the incoming matrix to the other processors (slaves) in the case where the matrix is centralized, and collect the solution. IMPLEMENTATION OF THE FORWARD MODELING To obtain insight regarding the advantages of the decomposition we consider a generic problem in airborne geophysics. The transmitter is a square loop, of 20 meters on a side, flown at 50 meters elevation. Three components of db/dt are acquired in the center of the loop. The waveform is a step-off and 35 time channels, equispaced in logarithmic time between 10 6 and 10 2 seconds, are measured. Data are collected every 20 meters along 20 east-west lines and there are 1000 transmitters. The acquisition geometry and the model, which is a conductive block in a resistive halfspace, are shown in Figure 1. In the numerical forward modeling the initial fields are the steady state fields and our time stepping begins at 10 6 seconds which is a decade before the first measurement time. Consequently, 35,000 solutions of Maxwell s equations (that is, 35 time steps and 1000 transmitters) are required to simulate the data. For a volume discretization we use a structured rectangular grid divided the earth into 70 3 cells. With our previous code that used iterative techniques the solution time for a single Maxwell system is about 2 minutes. The total time is therefore 2 35,000/N P where N P is the number of processors. In our case, where we use a dual quadcore INTEL XEON E5410 N P = 8 and so the forward modelling time, would be approximately 8750 minutes or 6 days. We now turn to the MUMPS factorization. Our equations are to be time stepped between times T 1 = 10 6 and T 2 = 10 2 seconds. The time step δt is determined by the smallest time needed and is µt 1 where µ is a constant. Here we use µ = 1 so the number of time steps is N δt = (T 2 T 1 )/δt = The factorization time, t f, depends upon the number of CPU s and the amount of available memory; for our configuration t f is 250s. With the factored system the time required to solve a forward problem, t s, is 0.6s. The total number of forward solutions is equal to N T X N δt where N T X is the number of transmitters. Thus the total time taken to solve our problem is T total = N T X N δt t s +t f. (9) The numbers are listed in the first row of Table I. Unfortunately the total time is still prohibitive. The difficulty arises because of the excessive number of time steps; keeping the same δt for the entire modeling requires 10 7 forward solutions. This difficulty can be circumvented by dividing our total time interval into subintervals each of which has a constant δt. One additional factorization is required for each subinterval.the exam-

3 Figure 1: The acquisition geometry and model. A conductive block of resistivity 1 Ω m is buried in a 100 Ω m halfspace. The top of the block is? meters below the surface. The locations of 1000 transmitters along east-west flight lines is shown. Data are obtained at the centers of each transmitter. ple here can be solved using 5 factorizations and 34 time steps. As per Table I, this reduces the total forward modeling time to 2.1E4s or about 5.6 hours. The simulated responses for the three components of db/dt are shown in Figure 2. N T X N P N δt t f t s N f T total E E4 Table 1: N T X is the number of transmitters; N p is the number of processors; N δt is the number of time steps; t f is the time required to factor the matrix; t s is the solution time for a forward problem once the matrix has been factorized; N f is the number of matrix factorization; T total is the time taken to solve the complete problem. Inversion The real benefit of using decomposition in our forward modeling becomes apparent when we treat the inverse problem. Our inversion algorithm (Haber et al, 2006), is based upon a Gauss- Newton procedure where a model m = lnσ is sought. The forward modeling is generically written as A(m)u = q where A is the forward modeling matrix, u are the fields, and q is the right hand side which contains the source terms. A is a large bidiagonal matrix of the form. Ã(m) 0 I Ã(m) A = (10) 0 I Ã(m) where Ã(m) is the discretized matrix in equation (7). At each Gauss-Newton iteration, the equation (J(m) T J(m) + βw T W)s = g(m) (11) Figure 2: The simulated responses for the three components of db/dt is solved to obtain a perturbation s from a current model m. W is a sparse regularization matrix, β is a constant, and g(m) is the gradient of the objective function. The sensitivity matrix is available as J(m) = QA(m) 1 G(m,u) where Q is an interpolation matrix that extracts the simulated data from the computed fields and G(m,u) is a known sparse matrix. G(m,u) = [A(m)u] m (12) We note that G requires the fields for all transmitters and for all times. Although these have been computed to evaluate the misfit and gradient, they need to be stored or recomputed to carry out the effect of multiplying J or J T on a vector. The system in equation (11) is solved using a preconditioned conjugate gradient (CG) solver. This requires that the matrix on the left be applied to a vector many times. The application of Jv or J v thus requires that a forward modeling be carried out. The implications of the factorization of the forward modeling matrix now become apparent. Firstly, the computation of the gradient requires a forward modeling. Secondly, even with a good preconditioner, N CG, the number of CG iterations needed to get a good solution, is Thirdly, once the perturbation s is calculated, a line search is required to find the scale factor α to update the model as m n+1 = m n + αs. If N CG = 15, the number of forward modelings is 31 plus any required for a line search. If there are multiple transmitters, then a complete forward modelling can involve thousands of solutions of Maxwell s equation and hence the matrix factorization

4 (a) Figure 3: The acquisition geometry and model. A conductive block of resistivity 1 Ω-m is buried in a 100 Ω-m halfspace. The top of the block is 8 meters below the surface. The locations of 9 transmitters are shown as well as the locations for the receivers. is highly important. As an example we invert data obtained from nine transmitter loops on the surface a flat earth. The data are three components of db/dt at an surface grid. The layout of the loops, the data locations, and the location of the buried block are shown in Figure 3. Data are recorded at 14 time channels between 10 5 and 10 3 seconds. The modelling time was divided so that three factorizations were required for a complete forward modelling. Gaussian noise was added to the 54,432 data prior to inversion. The inversion mesh as The time for a single factorization of the matrix was 6 seconds and the time for a complete forward modelling for all times and all transmitters was 28 seconds. Approximately 22 forward modelings were needed for each Gauss-Newton iteration and each iteration took approximately 10 minutes. XX iterations were required to successfully complete the inversion. A cross-sectional image of the recovered conductivity is shown in Figure 4. (b) Figure 4: The model recovered from inversion of the time domain data. (a) xy section. (b) xz section. ACKNOWLEDGMENTS This work was supported by NSERC CRDPJ and by MITEM (Multi-Source Inversion of Time Domain Electromagnetic Data) consortium partners: Newmont Mining Corp, CAMECO Corp, CRVD, Xstrata Plc, Teck Cominco Ltd, and Barrick Gold Corp. The second author was supported by NSF CMG grant xx and DOE grant yy.vvvvvvvvvvvvvvvlkjh CONCLUSIONS We have demonstrated the benefits of solving multi-source time domain electromagnetic problems by decomposing the matrix and using direct solvers. The advantages over traditional approaches that use iterative solvers depends upon the number of transmitters, the size of the discretized problem, the number of processors and the amount of memory available. Optimum use of this approach requires dividing the global time interval of interest into subintervals. If the factorization for each subinterval is stored, then Gauss-Newton iterations for the inverse problem can be efficiently carried out.

5 REFERENCES Amestoy, P. R., Guermouche, A., L Excellent, J.-Y., & Pralet, S Hybrid scheduling for the parallel solution of linear systems. Parallel Computing, 32, Dey, A., & Morrison, H. F Resistivity modeling for arbitrarily shaped three-dimensional structures. Geophysics, 44, Haber, E., & Ascher, U Fast finite volume simulation of 3d electromagnetic problems with highly discontinuous coefficients. SIAM Journal of Scientific Computing, 22, Haber, E., Ascher, U., Aruliah, D., & Oldenburg, D Fast simulation of 3d electromagnetic using potentials. Journal of Computational Physics, 163, Haber, E., Oldenburg, D., & Shekhtman, R Inversion of time domain 3d electromagnetic data. Geoph. J. Int. Vol 171, 2, pp

Computational methods for inverse problems

Computational methods for inverse problems Computational methods for inverse problems Eldad Haber contributors: Stefan Heldmann, Lauren Hanson, Lior Horesh, Jan Modersitzki, Uri Ascher, Doug Oldenburg Sponsored by: DOE and NSF Motivation and examples

More information

Three-dimensional inversion of borehole, time-domain, electromagnetic data for highly conductive ore-bodies.

Three-dimensional inversion of borehole, time-domain, electromagnetic data for highly conductive ore-bodies. Three-dimensional inversion of borehole, time-domain, electromagnetic data for highly conductive ore-bodies. Nigel Phillips, Doug Oldenburg, Eldad Haber, Roman Shekhtman. UBC-Geophysical Inversion Facility

More information

3D Inversion of Time-Domain Electromagnetic Data for Ground Water Aquifers

3D Inversion of Time-Domain Electromagnetic Data for Ground Water Aquifers 3D Inversion of Time-Domain Electromagnetic Data for Ground Water Aquifers Elliot M. Holtham 1, Mike McMillan 1 and Eldad Haber 2 (1) Computational Geosciences Inc. (2) University of British Columbia Summary

More information

E3D. A Program Library for Forward Modelling and Inversion of Frequency Domain EM Data over 3D Structures. Version 1.0

E3D. A Program Library for Forward Modelling and Inversion of Frequency Domain EM Data over 3D Structures. Version 1.0 E3D A Program Library for Forward Modelling and Inversion of Frequency Domain EM Data over 3D Structures Version 1.0 Developed under the consortium research project: COOPERATIVE INVERSION OF GEOPHYSICAL

More information

THE procedure used to solve inverse problems in areas such as Electrical Impedance

THE procedure used to solve inverse problems in areas such as Electrical Impedance 12TH INTL. CONFERENCE IN ELECTRICAL IMPEDANCE TOMOGRAPHY (EIT 2011), 4-6 MAY 2011, UNIV. OF BATH 1 Scaling the EIT Problem Alistair Boyle, Andy Adler, Andrea Borsic Abstract There are a number of interesting

More information

A direct solver with reutilization of previously-computed LU factorizations for h-adaptive finite element grids with point singularities

A direct solver with reutilization of previously-computed LU factorizations for h-adaptive finite element grids with point singularities Maciej Paszynski AGH University of Science and Technology Faculty of Computer Science, Electronics and Telecommunication Department of Computer Science al. A. Mickiewicza 30 30-059 Krakow, Poland tel.

More information

SCALABLE ALGORITHMS for solving large sparse linear systems of equations

SCALABLE ALGORITHMS for solving large sparse linear systems of equations SCALABLE ALGORITHMS for solving large sparse linear systems of equations CONTENTS Sparse direct solvers (multifrontal) Substructuring methods (hybrid solvers) Jacko Koster, Bergen Center for Computational

More information

EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI

EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI EFFICIENT SOLVER FOR LINEAR ALGEBRAIC EQUATIONS ON PARALLEL ARCHITECTURE USING MPI 1 Akshay N. Panajwar, 2 Prof.M.A.Shah Department of Computer Science and Engineering, Walchand College of Engineering,

More information

Parallel hp-finite Element Simulations of 3D Resistivity Logging Instruments

Parallel hp-finite Element Simulations of 3D Resistivity Logging Instruments Parallel hp-finite Element Simulations of 3D Resistivity Logging Instruments M. Paszyński 1,3, D. Pardo 1,2, L. Demkowicz 1, C. Torres-Verdin 2 1 Institute for Computational Engineering and Sciences 2

More information

Sparse Matrices Direct methods

Sparse Matrices Direct methods Sparse Matrices Direct methods Iain Duff STFC Rutherford Appleton Laboratory and CERFACS Summer School The 6th de Brùn Workshop. Linear Algebra and Matrix Theory: connections, applications and computations.

More information

DCIP3D. A Program Library for Forward Modelling and Inversion of DC/IP Data over 3D Structures. Version 5.0

DCIP3D. A Program Library for Forward Modelling and Inversion of DC/IP Data over 3D Structures. Version 5.0 DCIP3D A Program Library for Forward Modelling and Inversion of DC/IP Data over 3D Structures Version 50 Developed under the consortium research project: INVERSION OF 3D DC RESISTIVITY AND INDUCED POLARIZATION

More information

Practical aspects of running EH3D E. Haber Λ June 4, 2001 Abstract This document discuss of practical aspects of running the program EH3D. The goal of

Practical aspects of running EH3D E. Haber Λ June 4, 2001 Abstract This document discuss of practical aspects of running the program EH3D. The goal of Practical aspects of running EH3D E. Haber Λ June 4, Abstract This document discuss of practical aspects of running the program EH3D. The goal of this document is to help practitioners determine and tune

More information

Accelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method

Accelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method Accelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method Wenyong Pan 1, Kris Innanen 1 and Wenyuan Liao 2 1. CREWES Project, Department of Geoscience,

More information

H3DTD MUMPS. A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures.

H3DTD MUMPS. A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures. H3DTD MUMPS A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures. Version 1.6 Developed under the MITEM consortium Research Project Multi-Source

More information

A Parallel Implementation of the BDDC Method for Linear Elasticity

A Parallel Implementation of the BDDC Method for Linear Elasticity A Parallel Implementation of the BDDC Method for Linear Elasticity Jakub Šístek joint work with P. Burda, M. Čertíková, J. Mandel, J. Novotný, B. Sousedík Institute of Mathematics of the AS CR, Prague

More information

Approaches to Parallel Implementation of the BDDC Method

Approaches to Parallel Implementation of the BDDC Method Approaches to Parallel Implementation of the BDDC Method Jakub Šístek Includes joint work with P. Burda, M. Čertíková, J. Mandel, J. Novotný, B. Sousedík. Institute of Mathematics of the AS CR, Prague

More information

Model parametrization strategies for Newton-based acoustic full waveform

Model parametrization strategies for Newton-based acoustic full waveform Model parametrization strategies for Newton-based acoustic full waveform inversion Amsalu Y. Anagaw, University of Alberta, Edmonton, Canada, aanagaw@ualberta.ca Summary This paper studies the effects

More information

Study and implementation of computational methods for Differential Equations in heterogeneous systems. Asimina Vouronikoy - Eleni Zisiou

Study and implementation of computational methods for Differential Equations in heterogeneous systems. Asimina Vouronikoy - Eleni Zisiou Study and implementation of computational methods for Differential Equations in heterogeneous systems Asimina Vouronikoy - Eleni Zisiou Outline Introduction Review of related work Cyclic Reduction Algorithm

More information

Electromagnetic migration of marine CSEM data in areas with rough bathymetry Michael S. Zhdanov and Martin Čuma*, University of Utah

Electromagnetic migration of marine CSEM data in areas with rough bathymetry Michael S. Zhdanov and Martin Čuma*, University of Utah Electromagnetic migration of marine CSEM data in areas with rough bathymetry Michael S. Zhdanov and Martin Čuma*, University of Utah Summary In this paper we present a new approach to the interpretation

More information

Sparse Direct Solvers for Extreme-Scale Computing

Sparse Direct Solvers for Extreme-Scale Computing Sparse Direct Solvers for Extreme-Scale Computing Iain Duff Joint work with Florent Lopez and Jonathan Hogg STFC Rutherford Appleton Laboratory SIAM Conference on Computational Science and Engineering

More information

Parallel Implementations of Gaussian Elimination

Parallel Implementations of Gaussian Elimination s of Western Michigan University vasilije.perovic@wmich.edu January 27, 2012 CS 6260: in Parallel Linear systems of equations General form of a linear system of equations is given by a 11 x 1 + + a 1n

More information

Contents. I The Basic Framework for Stationary Problems 1

Contents. I The Basic Framework for Stationary Problems 1 page v Preface xiii I The Basic Framework for Stationary Problems 1 1 Some model PDEs 3 1.1 Laplace s equation; elliptic BVPs... 3 1.1.1 Physical experiments modeled by Laplace s equation... 5 1.2 Other

More information

Schemes for improving efficiency of pixel-based inversion algorithms for electromagnetic loggingwhile-drilling

Schemes for improving efficiency of pixel-based inversion algorithms for electromagnetic loggingwhile-drilling Schemes for improving efficiency of pixel-based inversion algorithms for electromagnetic loggingwhile-drilling measurements Y. Lin, A. Abubakar, T. M. Habashy, G. Pan, M. Li and V. Druskin, Schlumberger-Doll

More information

Parallel solution for finite element linear systems of. equations on workstation cluster *

Parallel solution for finite element linear systems of. equations on workstation cluster * Aug. 2009, Volume 6, No.8 (Serial No.57) Journal of Communication and Computer, ISSN 1548-7709, USA Parallel solution for finite element linear systems of equations on workstation cluster * FU Chao-jiang

More information

A fast direct solver for high frequency scattering from a cavity in two dimensions

A fast direct solver for high frequency scattering from a cavity in two dimensions 1/31 A fast direct solver for high frequency scattering from a cavity in two dimensions Jun Lai 1 Joint work with: Leslie Greengard (CIMS) Sivaram Ambikasaran (CIMS) Workshop on fast direct solver, Dartmouth

More information

Advanced Computer Graphics

Advanced Computer Graphics G22.2274 001, Fall 2009 Advanced Computer Graphics Project details and tools 1 Project Topics Computer Animation Geometric Modeling Computational Photography Image processing 2 Optimization All projects

More information

ME964 High Performance Computing for Engineering Applications

ME964 High Performance Computing for Engineering Applications ME964 High Performance Computing for Engineering Applications Outlining Midterm Projects Topic 3: GPU-based FEA Topic 4: GPU Direct Solver for Sparse Linear Algebra March 01, 2011 Dan Negrut, 2011 ME964

More information

A Random Variable Shape Parameter Strategy for Radial Basis Function Approximation Methods

A Random Variable Shape Parameter Strategy for Radial Basis Function Approximation Methods A Random Variable Shape Parameter Strategy for Radial Basis Function Approximation Methods Scott A. Sarra, Derek Sturgill Marshall University, Department of Mathematics, One John Marshall Drive, Huntington

More information

(Sparse) Linear Solvers

(Sparse) Linear Solvers (Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert

More information

Efficient 3D Gravity and Magnetic Modeling

Efficient 3D Gravity and Magnetic Modeling Efficient 3D Gravity and Magnetic Modeling X. Li Fugro Gravity & Magnetic Services Inc., Houston, Texas, USA Summary There are many different spatial-domain algorithms for 3D gravity and magnetic forward

More information

Null space computation of sparse singular matrices with MUMPS

Null space computation of sparse singular matrices with MUMPS Null space computation of sparse singular matrices with MUMPS Xavier Vasseur (CERFACS) In collaboration with Patrick Amestoy (INPT-IRIT, University of Toulouse and ENSEEIHT), Serge Gratton (INPT-IRIT,

More information

GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013

GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS. Kyle Spagnoli. Research EM Photonics 3/20/2013 GTC 2013: DEVELOPMENTS IN GPU-ACCELERATED SPARSE LINEAR ALGEBRA ALGORITHMS Kyle Spagnoli Research Engineer @ EM Photonics 3/20/2013 INTRODUCTION» Sparse systems» Iterative solvers» High level benchmarks»

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 20: Sparse Linear Systems; Direct Methods vs. Iterative Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 26

More information

Computational Fluid Dynamics - Incompressible Flows

Computational Fluid Dynamics - Incompressible Flows Computational Fluid Dynamics - Incompressible Flows March 25, 2008 Incompressible Flows Basis Functions Discrete Equations CFD - Incompressible Flows CFD is a Huge field Numerical Techniques for solving

More information

Using multifrontal hierarchically solver and HPC systems for 3D Helmholtz problem

Using multifrontal hierarchically solver and HPC systems for 3D Helmholtz problem Using multifrontal hierarchically solver and HPC systems for 3D Helmholtz problem Sergey Solovyev 1, Dmitry Vishnevsky 1, Hongwei Liu 2 Institute of Petroleum Geology and Geophysics SB RAS 1 EXPEC ARC,

More information

(x, y, z) m 2. (x, y, z) ...] T. m 2. m = [m 1. m 3. Φ = r T V 1 r + λ 1. m T Wm. m T L T Lm + λ 2. m T Hm + λ 3. t(x, y, z) = m 1

(x, y, z) m 2. (x, y, z) ...] T. m 2. m = [m 1. m 3. Φ = r T V 1 r + λ 1. m T Wm. m T L T Lm + λ 2. m T Hm + λ 3. t(x, y, z) = m 1 Class 1: Joint Geophysical Inversions Wed, December 1, 29 Invert multiple types of data residuals simultaneously Apply soft mutual constraints: empirical, physical, statistical Deal with data in the same

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

A Framework for the Upscaling of the Electrical Conductivity in the Quasi-static Maxwell s Equations

A Framework for the Upscaling of the Electrical Conductivity in the Quasi-static Maxwell s Equations A Framework for the Upscaling of the Electrical Conductivity in the Quasi-static Maxwell s Equations arxiv:1610.02948v1 [math.na] 7 Oct 2016 L. A. Caudillo-Mata*, E. Haber, L. J. Heagy, and C. Schwarzbach

More information

Inexact Full Newton Method for Full Waveform Inversion using Simultaneous encoded sources

Inexact Full Newton Method for Full Waveform Inversion using Simultaneous encoded sources Inexact Full Newton Method for Full Waveform Inversion using Simultaneous encoded sources Summary Amsalu Y. Anagaw, University of Alberta, Edmonton, Canada aanagaw@ualberta.ca and Mauricio D. Sacchi, University

More information

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear

More information

PROGRAMMING OF MULTIGRID METHODS

PROGRAMMING OF MULTIGRID METHODS PROGRAMMING OF MULTIGRID METHODS LONG CHEN In this note, we explain the implementation detail of multigrid methods. We will use the approach by space decomposition and subspace correction method; see Chapter:

More information

MUMPS. The MUMPS library. Abdou Guermouche and MUMPS team, June 22-24, Univ. Bordeaux 1 and INRIA

MUMPS. The MUMPS library. Abdou Guermouche and MUMPS team, June 22-24, Univ. Bordeaux 1 and INRIA The MUMPS library Abdou Guermouche and MUMPS team, Univ. Bordeaux 1 and INRIA June 22-24, 2010 MUMPS Outline MUMPS status Recently added features MUMPS and multicores? Memory issues GPU computing Future

More information

A projected Hessian matrix for full waveform inversion Yong Ma and Dave Hale, Center for Wave Phenomena, Colorado School of Mines

A projected Hessian matrix for full waveform inversion Yong Ma and Dave Hale, Center for Wave Phenomena, Colorado School of Mines A projected Hessian matrix for full waveform inversion Yong Ma and Dave Hale, Center for Wave Phenomena, Colorado School of Mines SUMMARY A Hessian matrix in full waveform inversion (FWI) is difficult

More information

A new take on FWI: Wavefield Reconstruction Inversion

A new take on FWI: Wavefield Reconstruction Inversion A new take on FWI: Wavefield Reconstruction Inversion T. van Leeuwen 1, F.J. Herrmann and B. Peters 1 Centrum Wiskunde & Informatica, Amsterdam, The Netherlands University of British Columbia, dept. of

More information

Efficient Finite Element Geometric Multigrid Solvers for Unstructured Grids on GPUs

Efficient Finite Element Geometric Multigrid Solvers for Unstructured Grids on GPUs Efficient Finite Element Geometric Multigrid Solvers for Unstructured Grids on GPUs Markus Geveler, Dirk Ribbrock, Dominik Göddeke, Peter Zajac, Stefan Turek Institut für Angewandte Mathematik TU Dortmund,

More information

Report of Linear Solver Implementation on GPU

Report of Linear Solver Implementation on GPU Report of Linear Solver Implementation on GPU XIANG LI Abstract As the development of technology and the linear equation solver is used in many aspects such as smart grid, aviation and chemical engineering,

More information

H3DTDinv MUMPS. A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures.

H3DTDinv MUMPS. A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures. H3DTDinv MUMPS A Program Library for Forward Modelling of Multi-Transmitter, Time-Domain Electromagnetic Data over 3D structures. Version 1.0 Developed under the MITEM consortium Research Project Multi-Source

More information

(Sparse) Linear Solvers

(Sparse) Linear Solvers (Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 1 Don t you just invert

More information

Development of a Maxwell Equation Solver for Application to Two Fluid Plasma Models. C. Aberle, A. Hakim, and U. Shumlak

Development of a Maxwell Equation Solver for Application to Two Fluid Plasma Models. C. Aberle, A. Hakim, and U. Shumlak Development of a Maxwell Equation Solver for Application to Two Fluid Plasma Models C. Aberle, A. Hakim, and U. Shumlak Aerospace and Astronautics University of Washington, Seattle American Physical Society

More information

Solving Partial Differential Equations on Overlapping Grids

Solving Partial Differential Equations on Overlapping Grids **FULL TITLE** ASP Conference Series, Vol. **VOLUME**, **YEAR OF PUBLICATION** **NAMES OF EDITORS** Solving Partial Differential Equations on Overlapping Grids William D. Henshaw Centre for Applied Scientific

More information

NEW ADVANCES IN GPU LINEAR ALGEBRA

NEW ADVANCES IN GPU LINEAR ALGEBRA GTC 2012: NEW ADVANCES IN GPU LINEAR ALGEBRA Kyle Spagnoli EM Photonics 5/16/2012 QUICK ABOUT US» HPC/GPU Consulting Firm» Specializations in:» Electromagnetics» Image Processing» Fluid Dynamics» Linear

More information

THE application of advanced computer architecture and

THE application of advanced computer architecture and 544 IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, VOL. 45, NO. 3, MARCH 1997 Scalable Solutions to Integral-Equation and Finite-Element Simulations Tom Cwik, Senior Member, IEEE, Daniel S. Katz, Member,

More information

smooth coefficients H. Köstler, U. Rüde

smooth coefficients H. Köstler, U. Rüde A robust multigrid solver for the optical flow problem with non- smooth coefficients H. Köstler, U. Rüde Overview Optical Flow Problem Data term and various regularizers A Robust Multigrid Solver Galerkin

More information

arxiv: v3 [math.na] 14 Jun 2018

arxiv: v3 [math.na] 14 Jun 2018 LAP: A LINEARIZE AND PROJECT METHOD FOR SOLVING INVERSE PROBLEMS WITH COUPLED VARIABLES JAMES L. HERRING, JAMES G. NAGY, AND LARS RUTHOTTO arxiv:1705.09992v3 [math.na] 14 Jun 2018 Abstract. Many inverse

More information

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND Student Submission for the 5 th OpenFOAM User Conference 2017, Wiesbaden - Germany: SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND TESSA UROIĆ Faculty of Mechanical Engineering and Naval Architecture, Ivana

More information

1.2 Numerical Solutions of Flow Problems

1.2 Numerical Solutions of Flow Problems 1.2 Numerical Solutions of Flow Problems DIFFERENTIAL EQUATIONS OF MOTION FOR A SIMPLIFIED FLOW PROBLEM Continuity equation for incompressible flow: 0 Momentum (Navier-Stokes) equations for a Newtonian

More information

CS 6210 Fall 2016 Bei Wang. Review Lecture What have we learnt in Scientific Computing?

CS 6210 Fall 2016 Bei Wang. Review Lecture What have we learnt in Scientific Computing? CS 6210 Fall 2016 Bei Wang Review Lecture What have we learnt in Scientific Computing? Let s recall the scientific computing pipeline observed phenomenon mathematical model discretization solution algorithm

More information

An Hybrid Approach for the Parallelization of a Block Iterative Algorithm

An Hybrid Approach for the Parallelization of a Block Iterative Algorithm An Hybrid Approach for the Parallelization of a Block Iterative Algorithm Carlos Balsa 1, Ronan Guivarch 2, Daniel Ruiz 2, and Mohamed Zenadi 2 1 CEsA FEUP, Porto, Portugal balsa@ipb.pt 2 Université de

More information

Downloaded 03/13/17 to Redistribution subject to SEG license or copyright; see Terms of Use at

Downloaded 03/13/17 to Redistribution subject to SEG license or copyright; see Terms of Use at Eldad Haber and Lars Ruthotto, Geophysical Inversion Facility, Department of Earth, Ocean and Atmospheric Sciences, University of British Columbia 2207 Main Mall, V6T 1Z4 Vancouver, British Columbia, Canada

More information

Lecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice

Lecture 11: Randomized Least-squares Approximation in Practice. 11 Randomized Least-squares Approximation in Practice Stat60/CS94: Randomized Algorithms for Matrices and Data Lecture 11-10/09/013 Lecture 11: Randomized Least-squares Approximation in Practice Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these

More information

Speedup Altair RADIOSS Solvers Using NVIDIA GPU

Speedup Altair RADIOSS Solvers Using NVIDIA GPU Innovation Intelligence Speedup Altair RADIOSS Solvers Using NVIDIA GPU Eric LEQUINIOU, HPC Director Hongwei Zhou, Senior Software Developer May 16, 2012 Innovation Intelligence ALTAIR OVERVIEW Altair

More information

Diffuse Optical Tomography, Inverse Problems, and Optimization. Mary Katherine Huffman. Undergraduate Research Fall 2011 Spring 2012

Diffuse Optical Tomography, Inverse Problems, and Optimization. Mary Katherine Huffman. Undergraduate Research Fall 2011 Spring 2012 Diffuse Optical Tomography, Inverse Problems, and Optimization Mary Katherine Huffman Undergraduate Research Fall 11 Spring 12 1. Introduction. This paper discusses research conducted in order to investigate

More information

GEOPHYS 242: Near Surface Geophysical Imaging. Class 8: Joint Geophysical Inversions Wed, April 20, 2011

GEOPHYS 242: Near Surface Geophysical Imaging. Class 8: Joint Geophysical Inversions Wed, April 20, 2011 GEOPHYS 4: Near Surface Geophysical Imaging Class 8: Joint Geophysical Inversions Wed, April, 11 Invert multiple types of data residuals simultaneously Apply soft mutual constraints: empirical, physical,

More information

Multi-azimuth velocity estimation

Multi-azimuth velocity estimation Stanford Exploration Project, Report 84, May 9, 2001, pages 1 87 Multi-azimuth velocity estimation Robert G. Clapp and Biondo Biondi 1 ABSTRACT It is well known that the inverse problem of estimating interval

More information

Intel MKL Sparse Solvers. Software Solutions Group - Developer Products Division

Intel MKL Sparse Solvers. Software Solutions Group - Developer Products Division Intel MKL Sparse Solvers - Agenda Overview Direct Solvers Introduction PARDISO: main features PARDISO: advanced functionality DSS Performance data Iterative Solvers Performance Data Reference Copyright

More information

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss

Robot Mapping. Least Squares Approach to SLAM. Cyrill Stachniss Robot Mapping Least Squares Approach to SLAM Cyrill Stachniss 1 Three Main SLAM Paradigms Kalman filter Particle filter Graphbased least squares approach to SLAM 2 Least Squares in General Approach for

More information

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General

Graphbased. Kalman filter. Particle filter. Three Main SLAM Paradigms. Robot Mapping. Least Squares Approach to SLAM. Least Squares in General Robot Mapping Three Main SLAM Paradigms Least Squares Approach to SLAM Kalman filter Particle filter Graphbased Cyrill Stachniss least squares approach to SLAM 1 2 Least Squares in General! Approach for

More information

Large-scale workflows for wave-equation based inversion in Julia

Large-scale workflows for wave-equation based inversion in Julia Large-scale workflows for wave-equation based inversion in Julia Philipp A. Witte, Mathias Louboutin and Felix J. Herrmann SLIM University of British Columbia Motivation Use Geophysics to understand the

More information

Sparse Multifrontal Performance Gains via NVIDIA GPU January 16, 2009

Sparse Multifrontal Performance Gains via NVIDIA GPU January 16, 2009 Sparse Multifrontal Performance Gains via NVIDIA GPU January 16, 2009 Dan l Pierce, PhD, MBA, CEO & President AAI Joint with: Yukai Hung, Chia-Chi Liu, Yao-Hung Tsai, Weichung Wang, and David Yu Access

More information

Towards a complete FEM-based simulation toolkit on GPUs: Geometric Multigrid solvers

Towards a complete FEM-based simulation toolkit on GPUs: Geometric Multigrid solvers Towards a complete FEM-based simulation toolkit on GPUs: Geometric Multigrid solvers Markus Geveler, Dirk Ribbrock, Dominik Göddeke, Peter Zajac, Stefan Turek Institut für Angewandte Mathematik TU Dortmund,

More information

Figure 6.1: Truss topology optimization diagram.

Figure 6.1: Truss topology optimization diagram. 6 Implementation 6.1 Outline This chapter shows the implementation details to optimize the truss, obtained in the ground structure approach, according to the formulation presented in previous chapters.

More information

Advanced Surface Based MoM Techniques for Packaging and Interconnect Analysis

Advanced Surface Based MoM Techniques for Packaging and Interconnect Analysis Electrical Interconnect and Packaging Advanced Surface Based MoM Techniques for Packaging and Interconnect Analysis Jason Morsey Barry Rubin, Lijun Jiang, Lon Eisenberg, Alina Deutsch Introduction Fast

More information

HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE

HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER PROF. BRYANT PROF. KAYVON 15618: PARALLEL COMPUTER ARCHITECTURE HYPERDRIVE IMPLEMENTATION AND ANALYSIS OF A PARALLEL, CONJUGATE GRADIENT LINEAR SOLVER AVISHA DHISLE PRERIT RODNEY ADHISLE PRODNEY 15618: PARALLEL COMPUTER ARCHITECTURE PROF. BRYANT PROF. KAYVON LET S

More information

AMS527: Numerical Analysis II

AMS527: Numerical Analysis II AMS527: Numerical Analysis II A Brief Overview of Finite Element Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao SUNY Stony Brook AMS527: Numerical Analysis II 1 / 25 Overview Basic concepts Mathematical

More information

Numerical methods for volume preserving image registration

Numerical methods for volume preserving image registration Numerical methods for volume preserving image registration SIAM meeting on imaging science 2004 Eldad Haber with J. Modersitzki haber@mathcs.emory.edu Emory University Volume preserving IR p.1/?? Outline

More information

1 Exercise: 1-D heat conduction with finite elements

1 Exercise: 1-D heat conduction with finite elements 1 Exercise: 1-D heat conduction with finite elements Reading This finite element example is based on Hughes (2000, sec. 1.1-1.15. 1.1 Implementation of the 1-D heat equation example In the previous two

More information

Super Matrix Solver-P-ICCG:

Super Matrix Solver-P-ICCG: Super Matrix Solver-P-ICCG: February 2011 VINAS Co., Ltd. Project Development Dept. URL: http://www.vinas.com All trademarks and trade names in this document are properties of their respective owners.

More information

2 Fundamentals of Serial Linear Algebra

2 Fundamentals of Serial Linear Algebra . Direct Solution of Linear Systems.. Gaussian Elimination.. LU Decomposition and FBS..3 Cholesky Decomposition..4 Multifrontal Methods. Iterative Solution of Linear Systems.. Jacobi Method Fundamentals

More information

MODELING MIXED BOUNDARY PROBLEMS WITH THE COMPLEX VARIABLE BOUNDARY ELEMENT METHOD (CVBEM) USING MATLAB AND MATHEMATICA

MODELING MIXED BOUNDARY PROBLEMS WITH THE COMPLEX VARIABLE BOUNDARY ELEMENT METHOD (CVBEM) USING MATLAB AND MATHEMATICA A. N. Johnson et al., Int. J. Comp. Meth. and Exp. Meas., Vol. 3, No. 3 (2015) 269 278 MODELING MIXED BOUNDARY PROBLEMS WITH THE COMPLEX VARIABLE BOUNDARY ELEMENT METHOD (CVBEM) USING MATLAB AND MATHEMATICA

More information

Efficient Imaging Algorithms on Many-Core Platforms

Efficient Imaging Algorithms on Many-Core Platforms Efficient Imaging Algorithms on Many-Core Platforms H. Köstler Dagstuhl, 22.11.2011 Contents Imaging Applications HDR Compression performance of PDE-based models Image Denoising performance of patch-based

More information

Space Filling Curves and Hierarchical Basis. Klaus Speer

Space Filling Curves and Hierarchical Basis. Klaus Speer Space Filling Curves and Hierarchical Basis Klaus Speer Abstract Real world phenomena can be best described using differential equations. After linearisation we have to deal with huge linear systems of

More information

Domain decomposition for 3D electromagnetic modeling

Domain decomposition for 3D electromagnetic modeling Earth Planets Space, 51, 1013 1018, 1999 Domain decomposition for 3D electromagnetic modeling Zonghou Xiong CRC AMET, Earth Science, Macquarie University, Sydney, NSW 2109, Australia (Received November

More information

ESPRESO ExaScale PaRallel FETI Solver. Hybrid FETI Solver Report

ESPRESO ExaScale PaRallel FETI Solver. Hybrid FETI Solver Report ESPRESO ExaScale PaRallel FETI Solver Hybrid FETI Solver Report Lubomir Riha, Tomas Brzobohaty IT4Innovations Outline HFETI theory from FETI to HFETI communication hiding and avoiding techniques our new

More information

Downloaded 18 Jul 2011 to Redistribution subject to SEG license or copyright; see Terms of Use at

Downloaded 18 Jul 2011 to Redistribution subject to SEG license or copyright; see Terms of Use at Rapid gravity and gravity gradiometry terrain correction via adaptive quadtree mesh discretization Kristofer Davis, M. Andy Kass, and Yaoguo Li, Center for Gravity, Electrical and Magnetic Studies, Colorado

More information

A substructure based parallel dynamic solution of large systems on homogeneous PC clusters

A substructure based parallel dynamic solution of large systems on homogeneous PC clusters CHALLENGE JOURNAL OF STRUCTURAL MECHANICS 1 (4) (2015) 156 160 A substructure based parallel dynamic solution of large systems on homogeneous PC clusters Semih Özmen, Tunç Bahçecioğlu, Özgür Kurç * Department

More information

Joint seismic traveltime and TEM inversion for near surface imaging Jide Nosakare Ogunbo*, Jie Zhang, GeoTomo LLC

Joint seismic traveltime and TEM inversion for near surface imaging Jide Nosakare Ogunbo*, Jie Zhang, GeoTomo LLC Jide Nosaare Ogunbo*, Jie Zhang, GeoTomo LLC Summary For a reliable interpretation of the subsurface structure, the joint geophysical inversion approach is becoming a viable tool. Seismic and EM methods

More information

26257 Nonlinear Inverse Modeling of Magnetic Anomalies due to Thin Sheets and Cylinders Using Occam s Method

26257 Nonlinear Inverse Modeling of Magnetic Anomalies due to Thin Sheets and Cylinders Using Occam s Method 26257 Nonlinear Inverse Modeling of Anomalies due to Thin Sheets and Cylinders Using Occam s Method R. Ghanati* (University of Tehran, Insitute of Geophysics), H.A. Ghari (University of Tehran, Insitute

More information

Lecture VIII. Global Approximation Methods: I

Lecture VIII. Global Approximation Methods: I Lecture VIII Global Approximation Methods: I Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Global Methods p. 1 /29 Global function approximation Global methods: function

More information

Scalable Algorithms in Optimization: Computational Experiments

Scalable Algorithms in Optimization: Computational Experiments Scalable Algorithms in Optimization: Computational Experiments Steven J. Benson, Lois McInnes, Jorge J. Moré, and Jason Sarich Mathematics and Computer Science Division, Argonne National Laboratory, Argonne,

More information

Mathematical Libraries and Application Software on JUQUEEN and JURECA

Mathematical Libraries and Application Software on JUQUEEN and JURECA Mitglied der Helmholtz-Gemeinschaft Mathematical Libraries and Application Software on JUQUEEN and JURECA JSC Training Course May 2017 I.Gutheil Outline General Informations Sequential Libraries Parallel

More information

High-Performance Computational Electromagnetic Modeling Using Low-Cost Parallel Computers

High-Performance Computational Electromagnetic Modeling Using Low-Cost Parallel Computers High-Performance Computational Electromagnetic Modeling Using Low-Cost Parallel Computers July 14, 1997 J Daniel S. Katz (Daniel.S.Katz@jpl.nasa.gov) Jet Propulsion Laboratory California Institute of Technology

More information

Lecture 15: More Iterative Ideas

Lecture 15: More Iterative Ideas Lecture 15: More Iterative Ideas David Bindel 15 Mar 2010 Logistics HW 2 due! Some notes on HW 2. Where we are / where we re going More iterative ideas. Intro to HW 3. More HW 2 notes See solution code!

More information

Analysis of the Out-of-Core Solution Phase of a Parallel Multifrontal Approach

Analysis of the Out-of-Core Solution Phase of a Parallel Multifrontal Approach Analysis of the Out-of-Core Solution Phase of a Parallel Multifrontal Approach P. Amestoy I.S. Duff A. Guermouche Tz. Slavova April 25, 200 Abstract We consider the parallel solution of sparse linear systems

More information

A comparison of Algorithms for Sparse Matrix. Real-time Multibody Dynamic Simulation

A comparison of Algorithms for Sparse Matrix. Real-time Multibody Dynamic Simulation A comparison of Algorithms for Sparse Matrix Factoring and Variable Reordering aimed at Real-time Multibody Dynamic Simulation Jose-Luis Torres-Moreno, Jose-Luis Blanco, Javier López-Martínez, Antonio

More information

Fast Algorithm for Matrix-Vector Multiply of Asymmetric Multilevel Block-Toeplitz Matrices

Fast Algorithm for Matrix-Vector Multiply of Asymmetric Multilevel Block-Toeplitz Matrices " Fast Algorithm for Matrix-Vector Multiply of Asymmetric Multilevel Block-Toeplitz Matrices B. E. Barrowes, F. L. Teixeira, and J. A. Kong Research Laboratory of Electronics, MIT, Cambridge, MA 02139-4307

More information

Efficient Iterative Semi-supervised Classification on Manifold

Efficient Iterative Semi-supervised Classification on Manifold . Efficient Iterative Semi-supervised Classification on Manifold... M. Farajtabar, H. R. Rabiee, A. Shaban, A. Soltani-Farani Sharif University of Technology, Tehran, Iran. Presented by Pooria Joulani

More information

Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics

Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics I. Pantle Fachgebiet Strömungsmaschinen Karlsruher Institut für Technologie KIT Motivation

More information

The determination of the correct

The determination of the correct SPECIAL High-performance SECTION: H i gh-performance computing computing MARK NOBLE, Mines ParisTech PHILIPPE THIERRY, Intel CEDRIC TAILLANDIER, CGGVeritas (formerly Mines ParisTech) HENRI CALANDRA, Total

More information

arxiv: v2 [cs.ms] 14 Dec 2016

arxiv: v2 [cs.ms] 14 Dec 2016 JINV A FLEXIBLE JULIA PACKAGE FOR PDE PARAMETER ESTIMATION LARS RUTHOTTO, ERAN TREISTER, AND ELDAD HABER arxiv:1606.07399v2 [cs.ms] 14 Dec 2016 Abstract. Estimating parameters of Partial Differential Equations

More information

Matrix-free IPM with GPU acceleration

Matrix-free IPM with GPU acceleration Matrix-free IPM with GPU acceleration Julian Hall, Edmund Smith and Jacek Gondzio School of Mathematics University of Edinburgh jajhall@ed.ac.uk 29th June 2011 Linear programming theory Primal-dual pair

More information