Exploring unstructured Poisson solvers for FDS

Similar documents
EXPLORING UNSTRUCTURED POISSON SOLVERS FOR FDS

Study and implementation of computational methods for Differential Equations in heterogeneous systems. Asimina Vouronikoy - Eleni Zisiou

Why Use the GPU? How to Exploit? New Hardware Features. Sparse Matrix Solvers on the GPU: Conjugate Gradients and Multigrid. Semiconductor trends

FOR P3: A monolithic multigrid FEM solver for fluid structure interaction

Contents. I The Basic Framework for Stationary Problems 1

Multigrid Solvers in CFD. David Emerson. Scientific Computing Department STFC Daresbury Laboratory Daresbury, Warrington, WA4 4AD, UK

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND

High Performance Computing: Tools and Applications

Performance of Implicit Solver Strategies on GPUs

Lecture 15: More Iterative Ideas

Parallel Computing. Slides credit: M. Quinn book (chapter 3 slides), A Grama book (chapter 3 slides)

CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman)

GPU Cluster Computing for FEM

Adaptive-Mesh-Refinement Pattern

Introduction to Multigrid and its Parallelization

Space Filling Curves and Hierarchical Basis. Klaus Speer

Dendro: Parallel algorithms for multigrid and AMR methods on 2:1 balanced octrees

Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics

Effects of Solvers on Finite Element Analysis in COMSOL MULTIPHYSICS

The Immersed Interface Method

(Sparse) Linear Solvers

Efficient Finite Element Geometric Multigrid Solvers for Unstructured Grids on GPUs

lecture 8 Groundwater Modelling -1

Flow and Heat Transfer in a Mixing Elbow

Semester Final Report

Multigrid Pattern. I. Problem. II. Driving Forces. III. Solution

Contents. F10: Parallel Sparse Matrix Computations. Parallel algorithms for sparse systems Ax = b. Discretized domain a metal sheet

1.2 Numerical Solutions of Flow Problems

Towards a complete FEM-based simulation toolkit on GPUs: Geometric Multigrid solvers

Ryian Hunter MAE 598

What is Multigrid? They have been extended to solve a wide variety of other problems, linear and nonlinear.

NIA CFD Seminar, October 4, 2011 Hyperbolic Seminar, NASA Langley, October 17, 2011

CFD Topology Optimization of Automotive Components

f xx + f yy = F (x, y)

Algebraic Multigrid (AMG) for Ground Water Flow and Oil Reservoir Simulation

Advanced Surface Based MoM Techniques for Packaging and Interconnect Analysis

Numerical Methods for PDEs. SSC Workgroup Meetings Juan J. Alonso October 8, SSC Working Group Meetings, JJA 1

CHAO YANG. Early Experience on Optimizations of Application Codes on the Sunway TaihuLight Supercomputer

Explicit and Implicit Coupling Strategies for Overset Grids. Jörg Brunswig, Manuel Manzke, Thomas Rung

Partial Differential Equations

Parallel Algorithms: Adaptive Mesh Refinement (AMR) method and its implementation

An Efficient, Geometric Multigrid Solver for the Anisotropic Diffusion Equation in Two and Three Dimensions

GPU Acceleration of Unmodified CSM and CFD Solvers

Fluent User Services Center

PROGRAMMING OF MULTIGRID METHODS

cuibm A GPU Accelerated Immersed Boundary Method

(Sparse) Linear Solvers

ESPRESO ExaScale PaRallel FETI Solver. Hybrid FETI Solver Report

Performance and accuracy of hardware-oriented. native-, solvers in FEM simulations

Driven Cavity Example

A COUPLED FINITE VOLUME SOLVER FOR THE SOLUTION OF LAMINAR TURBULENT INCOMPRESSIBLE AND COMPRESSIBLE FLOWS

Overview and Recent Developments of Dynamic Mesh Capabilities

AMS526: Numerical Analysis I (Numerical Linear Algebra)

A NURBS-BASED APPROACH FOR SHAPE AND TOPOLOGY OPTIMIZATION OF FLOW DOMAINS

High Performance Computing for PDE Some numerical aspects of Petascale Computing

Dendro: Parallel algorithms for multigrid and AMR methods on 2:1 balanced octrees

Recent developments for the multigrid scheme of the DLR TAU-Code

Using multifrontal hierarchically solver and HPC systems for 3D Helmholtz problem

Stream Function-Vorticity CFD Solver MAE 6263

TAU mesh deformation. Thomas Gerhold

THE MORTAR FINITE ELEMENT METHOD IN 2D: IMPLEMENTATION IN MATLAB

Numerical Modelling in Fortran: day 6. Paul Tackley, 2017

Matrix-free multi-gpu Implementation of Elliptic Solvers for strongly anisotropic PDEs

Investigation of cross flow over a circular cylinder at low Re using the Immersed Boundary Method (IBM)

Efficient O(N log N) algorithms for scattered data interpolation

Development of a Maxwell Equation Solver for Application to Two Fluid Plasma Models. C. Aberle, A. Hakim, and U. Shumlak

Automatic Generation of Algorithms and Data Structures for Geometric Multigrid. Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014

An adaptive discretization of incompressible flow using a multitude of moving Cartesian grids

Highly Parallel Multigrid Solvers for Multicore and Manycore Processors

HPC Algorithms and Applications

Parallel FEM Computation and Multilevel Graph Partitioning Xing Cai

AMath 483/583 Lecture 24. Notes: Notes: Steady state diffusion. Notes: Finite difference method. Outline:

AMath 483/583 Lecture 24

Numerical Methods to Solve 2-D and 3-D Elliptic Partial Differential Equations Using Matlab on the Cluster maya

CFD Best Practice Guidelines: A process to understand CFD results and establish Simulation versus Reality

Computational Fluid Dynamics - Incompressible Flows

Geometric Modeling Assignment 3: Discrete Differential Quantities

Smoothers. < interactive example > Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems

Massively Parallel Finite Element Simulations with deal.ii

Direct numerical simulation. in an annular pipe. of turbulent flow. Paolo Luchini & Maurizio Quadrio

Module 3 Mesh Generation

Numerical Algorithms

Performance Comparison between Blocking and Non-Blocking Communications for a Three-Dimensional Poisson Problem

T6: Position-Based Simulation Methods in Computer Graphics. Jan Bender Miles Macklin Matthias Müller

Block Distributed Schur Complement Preconditioners for CFD Computations on Many-Core Systems

PhD Student. Associate Professor, Co-Director, Center for Computational Earth and Environmental Science. Abdulrahman Manea.

LARGE-EDDY EDDY SIMULATION CODE FOR CITY SCALE ENVIRONMENTS

Efficient Multi-GPU CUDA Linear Solvers for OpenFOAM

Optimization of HOM Couplers using Time Domain Schemes

Numerical Implementation of Overlapping Balancing Domain Decomposition Methods on Unstructured Meshes

2 The Elliptic Test Problem

Directions: 1) Delete this text box 2) Insert desired picture here

Auto-tuning Multigrid with PetaBricks

Optimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH

AmgX 2.0: Scaling toward CORAL Joe Eaton, November 19, 2015

An adaptive discretization of incompressible flow using a multitude of moving Cartesian grids

Iterative Sparse Triangular Solves for Preconditioning

SHORTEST PATHS ON SURFACES GEODESICS IN HEAT

Data mining with sparse grids

Shape optimisation using breakthrough technologies

Transcription:

Exploring unstructured Poisson solvers for FDS Dr. Susanne Kilian hhpberlin - Ingenieure für Brandschutz 10245 Berlin - Germany

Agenda 1 Discretization of Poisson- Löser 2 Solvers for 3 Numerical Tests Conclusions 4 4

1 Discretization of the Structured versus unstructured Cartesian grids fire simulations Numerical

Pressure equation in FDS 1 Discretization of Elliptic partial differential equation of type Poisson + Boundary conditions Source terms of previous time step (radiation, combustion, etc.) must be solved at least twice per time step strongly coupled with velocity field

Finite difference discretization 1 Discretization of Discretization stencil in 2D: cell-centered specifies physical relations between single cells

Subdivision into meshes 1 Discretization of Single-Mesh: 1 global system of equations Multi-Mesh: M local systems of equations and are sparse matrices (only very few non-zeros entries)

Treatment of internal obstructions 1 Simple 2D-domain FDS velocity field Discretization of open outflow internal obstruction inflow

Structured Cartesian grids 1 Discretization of Gasphase and Solid -cells: uniform matrix stencils regardless of inner obstructions cells interior to obstructions are part of system of equations Matrix stencils don t care about obstructions

Structured Cartesian grids 1 Discretization of Advantages: very regular matrix structure (uniform numbering between neighboring cells) can be exploited efficiently in solution process (Example: FFT) Use of highly optimized solvers possible

Structured Cartesian grids 1 Discretization of Disadvantages: incorrect treatment of interior boundaries possible penetration of velocity field into internal solids

Structured Cartesian grids 1 Discretization of Disadvantages: incorrect treatment of interior boundaries possible penetration of velocity field into internal solids need of additional correction Losses of efficiency and accuracy

Unstructured Cartesian grids 1 Discretization of Only Gasphase -cells: individual matrix stencils by omitting internal obstructions cells interior to obstructions are not part of system of equations Matrix stencils care about obstructions

Unstructured Cartesian grids 1 Discretization of Advantages: correct setting of interior boundary conditions possible (homogeneous Neumann) less grid cells Higher accuracy, no additional correction

Unstructured Cartesian grids 1 Discretization of Disadvantages: loss of regular matrix structure (cells must store its neighbors) more general solvers needed (FFT doesn t work anymore) Application of optimized solvers difficult

Solvers for the Presentation of different strategies

Fast Fourier Transformation: FFT(tol) with velocity correction 2 Solvers for FFT(tol) Condition 1: Internal obstructions normal velocity components < tol Condition 2: Mesh interfaces Condition 1? Condition 2? difference of neighboring normal velocity components < tol FFT-solutions on single meshes are highly efficient and fast usable for structured grids only

Parallel LU-Decomposition: Cluster interface of Intel MKL Pardiso 2 Solvers for MKL - Init Initialization: first reordering of matrix structure then distributed LU-factorization MKL - Solve Pressure solution per time step: simple forward/backward substitution also praised to be very efficient usable for structured and unstructured grids

Scalable Recursive Clustering (ScaRC): Block-CG and -GMG Methods 2 Solvers for ScaRC-CG / ScaRC-GMG Preconditioning/Smoothing: Block-SSOR, Block-MKL Solution of coarse grid problem: Global CG, MKL Conjugate Gradient Methods (CG): solve equivalent minimization problem Geometric Multigrid Methods (GMG): use complete grid hierarchy with exact solution on coarsest grid level Meshwise strategies with 1 cell overlap reasonable convergence rates and scalability properties usable for structured and unstructured grids

Numerical tests Comparison of solvers on different geometries

Basic test geometries 3 Numerical Tests Cube - Cube without obstruction Cube + Cube with obstruction Cells per cube: 24 3, 48 3, 96 3, 192³, 240³, 288³ constant inflow of 1 m/s from the left, open outflow on the right comparison of structured FFT(tol) versus unstructured MKL und ScaRC

Different mesh decompositions 3 1-Mesh 8-Mesh 64-Mesh 1x1x1 2x2x2 4x4x4 Numerical Tests notations: Cube - (M) and Cube + (M) for corresponding M-mesh geometry comparison of all solvers on both geometries for M=1, 8, 64

Cube + (1): Velocity error 3 FFT(10-2 ) Ø 1 pressure iteration FFT(10-6 ) Ø 3,5 pressure iterations FFT(10-16 ) Ø 30 pressure iterations Numerical Tests 24 3 Cells, same simulation time and display range for all cases velocity correction successfully reduces error along internal obstructions number of pressure iterations increases if tolerance is driven to zero

Cube - (M) vs. Cube + (M): FFT(10-6 ) 3 Average of pressure iterations per time step for increasing M: 34 Numerical Tests Erfahrungen Geometry 48³ cells Number of meshes M 1 8 64 Cube - (M) 1 106 222 Cube + (M) 8 123 254 increasing number of pressure iterations if number of meshes is increased mesh decomposition causes higher rise than internal obstruction

Cube - (8) vs. Cube + (8): All solvers 3 Average time for 1 pressure solution: Numerical Tests 48 3 cells FFT(tol): extremely fast for coarse tol increasing costs for finer tol Time(s) MKL: best computing times (~ zero tol) ScaRC: good computing times (~ zero tol) FFT(10-2 ) FFT(10-4 ) FFT(10-6 ) MKL ScaRC

Cube + (8) vs. Cube + (64): All solvers 3 Average time for 1 pressure solution, growing problem size: Numerical Tests M=8, 48 3 cells M=64, 96 3 cells Time(s) Time(s) 24³ cells per mesh Method Method scalability gets worse if number of meshes is increased at constant load

Cube + (8): Costs MKL-method 3 Logarithmic scale!! Numerical Tests x 396 Storage x 164 High memory needs due to fill-in LU has much more non-zeros than A Number of non-zeros x 29 x 64 (FFT/ScaRC: very less memory needs) Runtime Expensive initialization Example: 8 Meshes with 96³ cells MKL-Init: MKL-Solve: ~ 5000 s 17 s 24³ 48³ 96³ 192³ Number of cells per mesh FFT and ScaRC can solve finer problems than MKL on given ressources (Example: FFT und ScaRC run for 288 3, MKL already fails for 240³)

Duct_Flow: Flow through a pipe 3 Case from FDS Verification Guide: 8 Meshes, 128³ cells 34 Numerical Tests Erfahrungen Method Average time for 1 pressure solution FFT(10-4 ) 41.3 s MKL 4.4 s ScaRC 7.5 s comparison of structured FFT(tol) versus unstructured MKL and ScaRC best times for MKL, reasonable times for ScaRC

Duct_Flow: Flow through a pipe 3 4 3 Numerical Tests FFT(10-4) MKL / ScaRC 8 Meshes, 64 3 Cells 8 Meshes, 64 3 Cells Erfahrungen FFT(tol): velocity correction slow (tol=10-4 needs ~1000 iterations) MKL / ScaRC: zero velocity error along pipe walls

Conclusions Summary and outlook

Summary and outlook 4 Summary no consistent overall picture yet, still more tests planned need to find a clever balance between: - accuracy (velocity error?) - performance (computational times for 1 Poisson solve?) - additional costs (storage, further libraries?) Conclusions 34 Erfahrungen Outlook test unstructured MKL and ScaRC: - to solve the implicit advection diffusion problem for scalars on the cut-cell region (IBM-method) - to solve the Laplace problem on the unstructured grid (as velocity correction) in combination with a structured FFT solution of the Poisson problem

hhpberlin Ingenieure für Brandschutz GmbH Thanks a lot for your attention Hauptsitz Rotherstraße 19 10245 Berlin Questions? Amtsgericht Berlin-Charlottenburg Register-Nr.: HRB 78 927 Ust-ID Nr.: DE217656065 Geschäftsführung: Dipl.-Ing. Karsten Foth Dipl.-Inf. BW [VWA] Stefan Truthän Beirat: Dipl.-Ing. Margot Ehrlicher Prof. Dr.-Ing. Dietmar Hosser Dr.-Ing. Karl-Heinz Schubert