Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm
|
|
- Stanley Manning
- 6 years ago
- Views:
Transcription
1 Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm March 17 March 21, 2014 Florian Schornbaum, Martin Bauer, Simon Bogner Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
2 Introductory Lecture Monday, March 17, 2014
3 Outline Introduction (of us / of participants) The walberla Framework What is the walberla Framework? Examples / Applications Course Schedule Introduction to the Lattice Boltzmann Method Introduction to the walberla Framework Data Structures / Underlying Concepts Domain Decomposition & Parallelization A Prototypical Simulation 3
4 Introduction (of us / of participants) Who are we? Who are you?
5 Introduction (of us / of participants) Martin Bauer [computational FAU] PhD LSS (chair for system simulation) core developer of walberla framework tutor for this compact course: Monday Friday Simon Bogner [math & computer FAU] PhD LSS (chair for system simulation) models and methods for lattice Boltzmann tutor for lectures on Tuesday & Wednesday Florian Schornbaum [computer FAU] PhD LSS (chair for system simulation) core developer of walberla framework tutor for this compact course: Monday Friday 5
6 Introduction (of us / of participants) Chair for System Simulation (1 st floor) Erlangen The University of Erlangen-Nuremberg [Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU)] has five faculties with over 270 chairs. 6
7 Introduction (of us / of participants) Chair for System Simulation (1 st floor) Faculty of Engineering (marked: Chair for System Simulation) FAU Erlangen-Nürnberg Faculty of Engineering Department of Computer Science Chair for System Simulation 7
8 Introduction (of us / of participants) Who are you?? Name? Field of study? Semester? Know-how in (computational) fluid mechanics? Programming Experience? (especially C++? C? Java?) 8
9 The walberla Framework What is the walberla Framework? Examples / Applications
10 The walberla Simulation Framework written in C++(11) main focus on CFD simulations based on the lattice Boltzmann method (LBM) but generally suitable for all kinds of numeric codes working with uniform domain decompositions (LSE solvers, phase field method) at its very core designed as an HPC software framework: scales from laptops to current petascale supercomputers largest simulation: 1,835,008 processes (IBM Blue Jülich) hybrid parallelization: MPI + OpenMP vectorization of compute kernels open source widely applicable lattice Boltzmann framework from Erlangen 10
11 The walberla Simulation Framework written in C++(11) main focus on CFD simulations based on the lattice Boltzmann method (LBM) but generally suitable for all kinds of numeric codes working with uniform domain decompositions (LSE solvers, phase field method) at its very core designed as an HPC software framework: scales from laptops to current petascale supercomputers largest simulation: 1,835,008 processes (IBM Blue Jülich) hybrid parallelization: MPI + OpenMP vectorization of compute kernels open source 11
12 The walberla Simulation Framework coupling with in-house rigid body physics engine pe automated build and test system: New git commit to central repository triggers builds on different systems with different compilers and different configurations (MPI on/off, OpenMP on/off, Debug/Release, float/double, ). After successful build, unit tests & set of applications are executed. support for different platforms (Linux, Windows) and compilers llvm/clang 12
13 The walberla Simulation Framework blood flow through coronary artery tree (C. Godenschwager) [ C. Godenschwager, F. Schornbaum, M. Bauer, H. Köstler, and U. Rüde, A Framework for Hybrid Parallel Flow Simulations with a Trillion Cells in Complex Geometries, SC13, Denver] domain decomposition of complex geometry 13
14 The walberla Simulation Framework electron beam melting metal powder: 3D printing (R. Ammer, M. Markl) [ liquid: walberla (LBM) metal powder: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 14
15 The walberla Simulation Framework electron beam melting metal powder: 3D printing (R. Ammer, M. Markl) [ liquid: walberla (LBM) metal powder: pe (fully resolved rigid bodies) ] camera footage of real application 15
16 The walberla Simulation Framework liquid-gas-solid flow simulation: stable floating positions of boxes with different densities (S. Bogner) [ liquid: walberla (LBM) boxes: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 16
17 The walberla Simulation Framework liquid-gas-solid flow simulation: bubble rising through pearls hovering under water (S. Bogner) [ liquid: walberla (LBM) pearls: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 17
18 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) THVZs 18
19 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) DNS (direct numerical simulation) Reynolds number: 1000 / D3Q19 TRT 4300 SuperMUC 101,466,432 fluid cells 25,800 blocks with 16 x 16 x 16 cells 19
20 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) number of different grid levels: time steps / sec (finest grid) total number of time steps (finest grid): 864,000 without refinement: 55.2 times more memory and 98.6 times the workload 20
21 Course Schedule
22 Course Schedule Monday Tuesday Wednesday Thursday Friday Introduction [Lecture] LB [Lecture] LB [Lecture] HPC (MPI) [Lecture] Project Presentations C++ [Lecture] C++ [Lecture] LB [Lecture] Work on Projects Project Presentations Lunch Break Lunch Break Lunch Break Lunch Break Lunch Break Getting to Know the Environment [Hands-On] walberla Tutorials [Hands-On] Introduction of Projects Work on Projects Work on Projects Work on Projects Morning time slots: and Lunch Break: Afternoon: (Friday: ) 22
23 Introduction to the Lattice Boltzmann Method
24 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Macroscopic quantities (velocity, density, ) can be calculated from the particle distribution functions. 24
25 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Macroscopic quantities (velocity, density, ) can be calculated from the particle distribution functions. 25
26 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Every particle distribution function is associated with a certain direction. 26
27 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 27
28 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 28
29 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 29
30 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] For the collision, different operators exist: SRT, TRT, MRT 30
31 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. These calculated values depend on the underlying boundary condition type. 31
32 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 32
33 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 33
34 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 34
35 Introduction to the LBM boundary treatment pre-streaming step boundary The lattice Boltzmann method allows to handle very complex geometries easily! Cells are typically just marked as being a boundary cell of a certain type: no-slip, free slip, velocity bounce-back, pressure, open boundary, [ flag field = special grid that stores these markings] 35
36 Introduction to the LBM A typical time step of the LBM looks like as follows: 1. boundary treatment is performed (= values in the boundary cells are prepared for streaming) 2. the streaming step is executed (requires nearest neighbor access) 3. the collision operator is evaluated (cell-local operation) A typical parallel time step includes communication: 1. all values of cells on the border of a process s partition of the domain are exchanged between neighboring processes (involves so-called ghost layer cells) 2. boundary treatment is performed (= values in the boundary cells are prepared for streaming) 3. the streaming step is executed (requires nearest neighbor access) 4. the collision operator is evaluated (cell-local operation) 36
37 Introduction to walberla Data Structures / Underlying Concepts Domain Decomposition & Parallelization A Prototypical Simulation
38 Introduction to walberla geometry given by surface mesh domain decomposition into blocks flow simulation only in here (example: complex geometry of an artery) load balancing empty blocks are discarded Load balancing can be based on either space-filling curves (Z-order/Morton order, Hilbert curve) using the underlying forest of octrees or graph partitioning (METIS, ). Whatever fits best the needs of the simulation. 38
39 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded Load balancing can be based on either space-filling curves (Z-order/Morton order, Hilbert curve) using the underlying forest of octrees or graph partitioning (METIS, ). Whatever fits best the needs of the simulation. 39
40 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded allocation of block data ( grids) The domain decomposition and load balancing can be performed during the actual simulation OR 40
41 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded separation of domain partitioning from simulation DISK DISK file size: kilobytes to few megabytes allocation of block data ( grids) 41
42 Introduction to walberla geometry given by surface mesh domain decomposition into blocks All of this (the entire pipeline) works just the same when grid refinement is used. load balancing empty blocks are discarded separation of domain partitioning from simulation DISK DISK file size: kilobytes to few megabytes allocation of block data ( grids) 42
43 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids grid refinement: octree-like decomposition special case of the much more general forest of octrees data structure non-uniform/refined grids forest of octrees: each block contains a uniform grid of the same size 2:1 balance between neighboring cells on level transitions (enables LB refinement algorithms) 43
44 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids special case of the much more general forest of octrees data structure non-uniform/refined grids octree-like decomposition of a globally uniform grid forest of octrees: each block contains a uniform grid with cells of the same size performance of uniform grids & good fitting to complex geometries 44
45 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids octree-like decomposition of the simulation space?????????????????????????????? special case of the much more general forest of octrees data structure non-uniform/refined grids forest of octrees: In general, each block contains whatever data was registered by the application: can be any arbitrary (sub)grid, scalar values, any object (in terms of C++), 45
46 Introduction to walberla Parallelization : data exchange on borders between blocks via ghost layers sender process receiver process (slightly more complicated for non-uniform domain decompositions, but the same general ideas still apply) typical parallelization of the underlying algorithm (e.g., LB): 1. perform communication: copy outermost layer of inner cells of perform communication: sender to ghost layer of receiver 2. perform one time step of the computation algorithm only on all inner cells of each block this computation algorithm is only allowed to access nearest neighbor cells during calculations! 46
47 Introduction to walberla A Prototypical Simulation: 1. construct domain decomposition (includes load balancing = distribution of blocks to all available processes) 2. add data to blocks (LB: at least one grid that stores the particle distribution functions and one grid that is used for marking cells as either being fluid or obstacle additional grids/classes may be required/useful [e.g., an object that is responsible for the boundary treatment]) 3. set-up geometry/boundaries (= mark cells as either being fluid or a boundary of a certain type information about the geometry/boundaries can be loaded from a mesh file, read in from a configuration file, set in the source code, etc.) 4. specify the algorithms/functions that are supposed to be executed in each time step (basic LB: communication, boundary treatment, stream & collide, output[optional]) 5. run the simulation = execute fixed number of time steps 47
48 THANK YOU FOR YOUR ATTENTION! QUESTIONS? (visit
simulation framework for piecewise regular grids
WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler
More informationPeta-Scale Simulations with the HPC Software Framework walberla:
Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,
More informationwalberla: Developing a Massively Parallel HPC Framework
walberla: Developing a Massively Parallel HPC Framework SIAM CS&E 2013, Boston February 26, 2013 Florian Schornbaum*, Christian Godenschwager*, Martin Bauer*, Matthias Markl, Ulrich Rüde* *Chair for System
More informationPerformance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla
Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla SIAM PP 2016, April 13 th 2016 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer,
More informationMassively Parallel Phase Field Simulations using HPC Framework walberla
Massively Parallel Phase Field Simulations using HPC Framework walberla SIAM CSE 2015, March 15 th 2015 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer, Harald Köstler and Ulrich
More informationThe walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms
The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms Harald Köstler, Uli Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationA Python extension for the massively parallel framework walberla
A Python extension for the massively parallel framework walberla PyHPC at SC 14, November 17 th 2014 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler
More information(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more
Parallel Free-Surface Extension of the Lattice-Boltzmann Method A Lattice-Boltzmann Approach for Simulation of Two-Phase Flows Stefan Donath (LSS Erlangen, stefan.donath@informatik.uni-erlangen.de) Simon
More informationSimulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method
Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method June 21, 2011 Introduction Free Surface LBM Liquid-Gas-Solid Flows Parallel Computing Examples and More References Fig. Simulation
More informationLattice Boltzmann Methods on the way to exascale
Lattice Boltzmann Methods on the way to exascale Ulrich Rüde (LSS Erlangen, ulrich.ruede@fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de HIGH PERFORMANCE
More informationSoftware and Performance Engineering for numerical codes on GPU clusters
Software and Performance Engineering for numerical codes on GPU clusters H. Köstler International Workshop of GPU Solutions to Multiscale Problems in Science and Engineering Harbin, China 28.7.2010 2 3
More informationSimulation of moving Particles in 3D with the Lattice Boltzmann Method
Simulation of moving Particles in 3D with the Lattice Boltzmann Method, Nils Thürey, Christian Feichtinger, Hans-Joachim Schmid Chair for System Simulation University Erlangen/Nuremberg Chair for Particle
More informationPerformance Analysis of the Lattice Boltzmann Method on x86-64 Architectures
Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Jan Treibig, Simon Hausmann, Ulrich Ruede Zusammenfassung The Lattice Boltzmann method (LBM) is a well established algorithm
More informationLattice Boltzmann with CUDA
Lattice Boltzmann with CUDA Lan Shi, Li Yi & Liyuan Zhang Hauptseminar: Multicore Architectures and Programming Page 1 Outline Overview of LBM An usage of LBM Algorithm Implementation in CUDA and Optimization
More informationA Contact Angle Model for the Parallel Free Surface Lattice Boltzmann Method in walberla Stefan Donath (stefan.donath@informatik.uni-erlangen.de) Computer Science 10 (System Simulation) University of Erlangen-Nuremberg
More informationInternational Supercomputing Conference 2009
International Supercomputing Conference 2009 Implementation of a Lattice-Boltzmann-Method for Numerical Fluid Mechanics Using the nvidia CUDA Technology E. Riegel, T. Indinger, N.A. Adams Technische Universität
More informationLATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS
14 th European Conference on Mixing Warszawa, 10-13 September 2012 LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS Felix Muggli a, Laurent Chatagny a, Jonas Lätt b a Sulzer Markets & Technology
More informationReconstruction of Trees from Laser Scan Data and further Simulation Topics
Reconstruction of Trees from Laser Scan Data and further Simulation Topics Helmholtz-Research Center, Munich Daniel Ritter http://www10.informatik.uni-erlangen.de Overview 1. Introduction of the Chair
More informationHigh Performance Computing
High Performance Computing ADVANCED SCIENTIFIC COMPUTING Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich
More informationFRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)
FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) walberla: Visualization of Fluid
More informationFrom Notebooks to Supercomputers: Tap the Full Potential of Your CUDA Resources with LibGeoDecomp
From Notebooks to Supercomputers: Tap the Full Potential of Your CUDA Resources with andreas.schaefer@cs.fau.de Friedrich-Alexander-Universität Erlangen-Nürnberg GPU Technology Conference 2013, San José,
More informationNavier-Stokes & Flow Simulation
Last Time? Navier-Stokes & Flow Simulation Pop Worksheet! Teams of 2. Hand in to Jeramey after we discuss. Sketch the first few frames of a 2D explicit Euler mass-spring simulation for a 2x3 cloth network
More informationPerformance and Software-Engineering Considerations for Massively Parallel Simulations
Performance and Software-Engineering Considerations for Massively Parallel Simulations Ulrich Rüde (ruede@cs.fau.de) Ben Bergen, Frank Hülsemann, Christoph Freundl Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationNumerical Algorithms on Multi-GPU Architectures
Numerical Algorithms on Multi-GPU Architectures Dr.-Ing. Harald Köstler 2 nd International Workshops on Advances in Computational Mechanics Yokohama, Japan 30.3.2010 2 3 Contents Motivation: Applications
More informationIntroducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method
Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method G. Wellein, T. Zeiser, G. Hager HPC Services Regional Computing Center A. Nitsure, K. Iglberger, U. Rüde Chair for System
More informationParallel Simulation of Dendritic Growth On Unstructured Grids
Parallel Simulation of Dendritic Growth On Unstructured Grids, Julian Hammer, Dietmar Fey Friedrich-Alexander-Universität Erlangen-Nürnberg IA 3 Workshop @SC11, Nov. 13th, 2011 Outline 1 What and why?
More informationTechnical Report TR
Technical Report TR-2015-09 Boundary condition enforcing methods for smoothed particle hydrodynamics Arman Pazouki 1, Baofang Song 2, Dan Negrut 1 1 University of Wisconsin-Madison, Madison, WI, 53706-1572,
More informationLattice Boltzmann methods on the way to exascale
Lattice Boltzmann methods on the way to exascale Ulrich Rüde LSS Erlangen and CERFACS Toulouse ulrich.ruede@fau.de Centre Européen de Recherche et de Formation Avancée en Calcul Scientifique www.cerfacs.fr
More informationNavier-Stokes & Flow Simulation
Last Time? Navier-Stokes & Flow Simulation Implicit Surfaces Marching Cubes/Tetras Collision Detection & Response Conservative Bounding Regions backtracking fixing Today Flow Simulations in Graphics Flow
More informationCS GPU and GPGPU Programming Lecture 2: Introduction; GPU Architecture 1. Markus Hadwiger, KAUST
CS 380 - GPU and GPGPU Programming Lecture 2: Introduction; GPU Architecture 1 Markus Hadwiger, KAUST Reading Assignment #2 (until Feb. 17) Read (required): GLSL book, chapter 4 (The OpenGL Programmable
More informationLARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS
ECCOMAS Congress 2016 VII European Congress on Computational Methods in Applied Sciences and Engineering M. Papadrakakis, V. Papadopoulos, G. Stefanou, V. Plevris (eds.) Crete Island, Greece, 5 10 June
More informationTowards PetaScale Computational Science
Towards PetaScale Computational Science U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with many Lehrstuhl für Informatik 10 (Systemsimulation) Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationNavier-Stokes & Flow Simulation
Last Time? Navier-Stokes & Flow Simulation Optional Reading for Last Time: Spring-Mass Systems Numerical Integration (Euler, Midpoint, Runge-Kutta) Modeling string, hair, & cloth HW2: Cloth & Fluid Simulation
More informationLarge scale Imaging on Current Many- Core Platforms
Large scale Imaging on Current Many- Core Platforms SIAM Conf. on Imaging Science 2012 May 20, 2012 Dr. Harald Köstler Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen,
More informationDomain Decomposition for Colloid Clusters. Pedro Fernando Gómez Fernández
Domain Decomposition for Colloid Clusters Pedro Fernando Gómez Fernández MSc in High Performance Computing The University of Edinburgh Year of Presentation: 2004 Authorship declaration I, Pedro Fernando
More informationChallenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids
Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Sebastian Kuckuk FAU Erlangen-Nürnberg 18.01.2016 HiStencils 2016, Prague, Czech Republic Outline Outline Scope
More informationCMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman)
CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC Guest Lecturer: Sukhyun Song (original slides by Alan Sussman) Parallel Programming with Message Passing and Directives 2 MPI + OpenMP Some applications can
More informationCAF versus MPI Applicability of Coarray Fortran to a Flow Solver
CAF versus MPI Applicability of Coarray Fortran to a Flow Solver Manuel Hasert, Harald Klimach, Sabine Roller m.hasert@grs-sim.de Applied Supercomputing in Engineering Motivation We develop several CFD
More informationJoint Advanced Student School 2007 Martin Dummer
Sierpiński-Curves Joint Advanced Student School 2007 Martin Dummer Statement of the Problem What is the best way to store a triangle mesh efficiently in memory? The following points are desired : Easy
More informationA Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla
A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler, and Ulrich
More informationSPEED-UP GEARBOX SIMULATIONS BY INTEGRATING SCORG. Dr. Christine Klier, Sahand Saheb-Jahromi, Ludwig Berger*
SPEED-UP GEARBOX SIMULATIONS BY INTEGRATING SCORG Dr. Christine Klier, Sahand Saheb-Jahromi, Ludwig Berger* CFD SCHUCK ENGINEERING Engineering Services in computational fluid Dynamics (CFD) 25 employees
More informationHigh-Performance Scientific Computing
High-Performance Scientific Computing Instructor: Randy LeVeque TA: Grady Lemoine Applied Mathematics 483/583, Spring 2011 http://www.amath.washington.edu/~rjl/am583 World s fastest computers http://top500.org
More informationMassively Parallel Finite Element Simulations with deal.ii
Massively Parallel Finite Element Simulations with deal.ii Timo Heister, Texas A&M University 2012-02-16 SIAM PP2012 joint work with: Wolfgang Bangerth, Carsten Burstedde, Thomas Geenen, Martin Kronbichler
More informationTwo-Phase flows on massively parallel multi-gpu clusters
Two-Phase flows on massively parallel multi-gpu clusters Peter Zaspel Michael Griebel Institute for Numerical Simulation Rheinische Friedrich-Wilhelms-Universität Bonn Workshop Programming of Heterogeneous
More informationSustainability and Efficiency for Simulation Software in the Exascale Era
Sustainability and Efficiency for Simulation Software in the Exascale Era Dominik Thönnes, Ulrich Rüde, Nils Kohl Chair for System Simulation, University of Erlangen-Nürnberg March 09, 2018 SIAM Conference
More informationFree Surface Lattice-Boltzmann fluid simulations. with and without level sets.
ree Surface Lattice-Boltzmann fluid simulations with and without level sets Nils Thürey, Ulrich Rüde University of Erlangen-Nuremberg System Simulation roup Cauerstr. 6, 91054 Erlangen, ermany Email: Nils.Thuerey@cs.fau.de
More informationPerformance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures
Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Dirk Ribbrock, Markus Geveler, Dominik Göddeke, Stefan Turek Angewandte Mathematik, Technische Universität Dortmund
More informationWebinar #3 Lattice Boltzmann method for CompBioMed (incl. Palabos)
Webinar series A Centre of Excellence in Computational Biomedicine Webinar #3 Lattice Boltzmann method for CompBioMed (incl. Palabos) 19 March 2018 The webinar will start at 12pm CET / 11am GMT Dr Jonas
More informationHIGH PERFORMANCE COMPUTATION (HPC) FOR THE
HIGH PERFORMANCE COMPUTATION (HPC) FOR THE DEVELOPMENT OF FLUIDIZED BED TECHNOLOGIES FOR BIOMASS GASIFICATION AND CO2 CAPTURE P. Fede, H. Neau, O. Simonin Université de Toulouse; INPT, UPS ; IMFT ; 31400
More informationArchitecture Aware Multigrid
Architecture Aware Multigrid U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with D. Ritter, T. Gradl, M. Stürmer, H. Köstler, J. Treibig and many more students Lehrstuhl für Informatik 10 (Systemsimulation)
More informationVisualization. Images are used to aid in understanding of data. Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [chapter 26]
Visualization Images are used to aid in understanding of data Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [chapter 26] Tumor SCI, Utah Scientific Visualization Visualize large
More informationIntroduction CPS343. Spring Parallel and High Performance Computing. CPS343 (Parallel and HPC) Introduction Spring / 29
Introduction CPS343 Parallel and High Performance Computing Spring 2018 CPS343 (Parallel and HPC) Introduction Spring 2018 1 / 29 Outline 1 Preface Course Details Course Requirements 2 Background Definitions
More informationCAF versus MPI - Applicability of Coarray Fortran to a Flow Solver
CAF versus MPI - Applicability of Coarray Fortran to a Flow Solver Manuel Hasert, Harald Klimach, Sabine Roller 2 German Research School for Simulation Sciences GmbH, 52062 Aachen, Germany 2 RWTH Aachen
More informationDynamic Mode Decomposition analysis of flow fields from CFD Simulations
Dynamic Mode Decomposition analysis of flow fields from CFD Simulations Technische Universität München Thomas Indinger Lukas Haag, Daiki Matsumoto, Christoph Niedermeier in collaboration with Agenda Motivation
More informationSIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS
XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22, 2012 SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC
More informationTutorial 17. Using the Mixture and Eulerian Multiphase Models
Tutorial 17. Using the Mixture and Eulerian Multiphase Models Introduction: This tutorial examines the flow of water and air in a tee junction. First you will solve the problem using the less computationally-intensive
More informationBOUNDLESS FLUIDS USING THE LATTICE-BOLTZMANN METHOD. A Thesis. Presented to. the Faculty of California Polytechnic State University.
BOUNDLESS FLUIDS USING THE LATTICE-BOLTZMANN METHOD A Thesis Presented to the Faculty of California Polytechnic State University San Luis Obispo In Partial Fulfillment of the Requirements for the Degree
More informationForest-of-octrees AMR: algorithms and interfaces
Forest-of-octrees AMR: algorithms and interfaces Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac, Georg Stadler, Lucas C. Wilcox Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität
More informationExpress Introductory Training in ANSYS Fluent Workshop 02 Using the Discrete Phase Model (DPM)
Express Introductory Training in ANSYS Fluent Workshop 02 Using the Discrete Phase Model (DPM) Dimitrios Sofialidis Technical Manager, SimTec Ltd. Mechanical Engineer, PhD PRACE Autumn School 2013 - Industry
More informationNVIDIA. Interacting with Particle Simulation in Maya using CUDA & Maximus. Wil Braithwaite NVIDIA Applied Engineering Digital Film
NVIDIA Interacting with Particle Simulation in Maya using CUDA & Maximus Wil Braithwaite NVIDIA Applied Engineering Digital Film Some particle milestones FX Rendering Physics 1982 - First CG particle FX
More informationCUDA. Fluid simulation Lattice Boltzmann Models Cellular Automata
CUDA Fluid simulation Lattice Boltzmann Models Cellular Automata Please excuse my layout of slides for the remaining part of the talk! Fluid Simulation Navier Stokes equations for incompressible fluids
More informationGPU-based Distributed Behavior Models with CUDA
GPU-based Distributed Behavior Models with CUDA Courtesy: YouTube, ISIS Lab, Universita degli Studi di Salerno Bradly Alicea Introduction Flocking: Reynolds boids algorithm. * models simple local behaviors
More informationRealtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research
1 Realtime Water Simulation on GPU Nuttapong Chentanez NVIDIA Research 2 3 Overview Approaches to realtime water simulation Hybrid shallow water solver + particles Hybrid 3D tall cell water solver + particles
More informationAsynchronous OpenCL/MPI numerical simulations of conservation laws
Asynchronous OpenCL/MPI numerical simulations of conservation laws Philippe HELLUY 1,3, Thomas STRUB 2. 1 IRMA, Université de Strasbourg, 2 AxesSim, 3 Inria Tonus, France IWOCL 2015, Stanford Conservation
More informationSupersonic Flow Over a Wedge
SPC 407 Supersonic & Hypersonic Fluid Dynamics Ansys Fluent Tutorial 2 Supersonic Flow Over a Wedge Ahmed M Nagib Elmekawy, PhD, P.E. Problem Specification A uniform supersonic stream encounters a wedge
More informationHigh Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters
SIAM PP 2014 High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters C. Riesinger, A. Bakhtiari, M. Schreiber Technische Universität München February 20, 2014
More informationHeight Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 12] April 23, 2002 Frank Pfenning Carnegie Mellon University
15-462 Computer Graphics I Lecture 21 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 12] April 23, 2002 Frank Pfenning Carnegie Mellon University http://www.cs.cmu.edu/~fp/courses/graphics/
More informationIntroduction to the immersed boundary method
Introduction to the immersed boundary method Motivation. Hydrodynamics and boundary conditions The incompressible Navier-Stokes equations, ( ) u ρ + (u )u = p + ρν 2 u + f, () t are partial differential
More informationAdaptive Refinement Tree (ART) code. N-Body: Parallelization using OpenMP and MPI
Adaptive Refinement Tree (ART) code N-Body: Parallelization using OpenMP and MPI 1 Family of codes N-body: OpenMp N-body: MPI+OpenMP N-body+hydro+cooling+SF: OpenMP N-body+hydro+cooling+SF: MPI 2 History:
More informationPHYSICALLY BASED ANIMATION
PHYSICALLY BASED ANIMATION CS148 Introduction to Computer Graphics and Imaging David Hyde August 2 nd, 2016 WHAT IS PHYSICS? the study of everything? WHAT IS COMPUTATION? the study of everything? OUTLINE
More informationLATTICE-BOLTZMANN AND COMPUTATIONAL FLUID DYNAMICS
LATTICE-BOLTZMANN AND COMPUTATIONAL FLUID DYNAMICS NAVIER-STOKES EQUATIONS u t + u u + 1 ρ p = Ԧg + ν u u=0 WHAT IS COMPUTATIONAL FLUID DYNAMICS? Branch of Fluid Dynamics which uses computer power to approximate
More informationFree Surface Flows with Moving and Deforming Objects for LBM
Free Surface Flows with Moving and Deforming Objects for LBM Nils Thürey, Klaus Iglberger, Ulrich Rüde Institute for System Simulation (LSS), University of Erlangen-Nuremberg, Cauerstr. 6, 91058 Erlangen,
More informationAdarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs)
OBJECTIVE FLUID SIMULATIONS Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) The basic objective of the project is the implementation of the paper Stable Fluids (Jos Stam, SIGGRAPH 99). The final
More informationEnzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy
Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2
More informationNIA CFD Futures Conference Hampton, VA; August 2012
Petascale Computing and Similarity Scaling in Turbulence P. K. Yeung Schools of AE, CSE, ME Georgia Tech pk.yeung@ae.gatech.edu NIA CFD Futures Conference Hampton, VA; August 2012 10 2 10 1 10 4 10 5 Supported
More informationShape of Things to Come: Next-Gen Physics Deep Dive
Shape of Things to Come: Next-Gen Physics Deep Dive Jean Pierre Bordes NVIDIA Corporation Free PhysX on CUDA PhysX by NVIDIA since March 2008 PhysX on CUDA available: August 2008 GPU PhysX in Games Physical
More informationDirections: 1) Delete this text box 2) Insert desired picture here
Directions: 1) Delete this text box 2) Insert desired picture here Multi-Disciplinary Applications using Overset Grid Technology in STAR-CCM+ CD-adapco Dmitry Pinaev, Frank Schäfer, Eberhard Schreck Outline
More informationPreliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede
Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Qingyu Meng, Alan Humphrey, John Schmidt, Martin Berzins Thanks to: TACC Team for early access to Stampede J. Davison
More informationTechnical Tricks of Coarse-Grained MD Visualization with VMD
Technical Tricks of Coarse-Grained MD Visualization with VMD Institut für Computerphysik, Universität Stuttgart Stuttgart, Germany Arbitrary Units Lennard-Jones-Potential All-atom MD simulations Lengths
More informationA Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications
A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference
More informationGeneric finite element capabilities for forest-of-octrees AMR
Generic finite element capabilities for forest-of-octrees AMR Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität
More information11/1/13. Visualization. Scientific Visualization. Types of Data. Height Field. Contour Curves. Meshes
CSCI 420 Computer Graphics Lecture 26 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 2.11] Jernej Barbic University of Southern California Scientific Visualization
More informationVisualization. CSCI 420 Computer Graphics Lecture 26
CSCI 420 Computer Graphics Lecture 26 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 11] Jernej Barbic University of Southern California 1 Scientific Visualization
More informationOptimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH
Optimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH Carsten Clauss, Martin Pöppe, Thomas Bemmerl carsten@lfbs.rwth-aachen.de http://www.mp-mpich.de Lehrstuhl für Betriebssysteme
More informationRAPID LARGE-SCALE CARTESIAN MESHING FOR AERODYNAMIC COMPUTATIONS
RAPID LARGE-SCALE CARTESIAN MESHING FOR AERODYNAMIC COMPUTATIONS Daisuke Sasaki*, Kazuhiro Nakahashi** *Department of Aeronautics, Kanazawa Institute of Technology, **JAXA Keywords: Meshing, Cartesian
More informationAccurate and Efficient Turbomachinery Simulation. Chad Custer, PhD Turbomachinery Technical Specialist
Accurate and Efficient Turbomachinery Simulation Chad Custer, PhD Turbomachinery Technical Specialist Outline Turbomachinery simulation advantages Axial fan optimization Description of design objectives
More informationAdvanced Parallel Programming
Sebastian von Alfthan Jussi Enkovaara Pekka Manninen Advanced Parallel Programming February 15-17, 2016 PRACE Advanced Training Center CSC IT Center for Science Ltd, Finland All material (C) 2011-2016
More informationMASSIVELY-PARALLEL MULTI-GPU SIMULATIONS FOR FAST AND ACCURATE AUTOMOTIVE AERODYNAMICS
6th European Conference on Computational Mechanics (ECCM 6) 7th European Conference on Computational Fluid Dynamics (ECFD 7) -5 June 28, Glasgow, UK MASSIVELY-PARALLEL MULTI-GPU SIMULATIONS FOR FAST AND
More informationCS 231. Fluid simulation
CS 231 Fluid simulation Why Simulate Fluids? Feature film special effects Computer games Medicine (e.g. blood flow in heart) Because it s fun Fluid Simulation Called Computational Fluid Dynamics (CFD)
More informationParallel Mesh Multiplication for Code_Saturne
Parallel Mesh Multiplication for Code_Saturne Pavla Kabelikova, Ales Ronovsky, Vit Vondrak a Dept. of Applied Mathematics, VSB-Technical University of Ostrava, Tr. 17. listopadu 15, 708 00 Ostrava, Czech
More informationAlex Li 11/20/2009. Chris Wojtan, Nils Thurey, Markus Gross, Greg Turk
Alex Li 11/20/2009 Chris Wojtan, Nils Thurey, Markus Gross, Greg Turk duction Overview of Lagrangian of Topological s Altering the Topology 2 Presents a method for accurately tracking the moving surface
More informationSTAR Global Conference 2012 Noordwijk, March 20-21, The InDesA Virtual Test Bench. Dr. Fabiano Bet Dr. Gerald Seider
STAR Global Conference 2012 Noordwijk, March 20-21, 2012 The InDesA Dr. Fabiano Bet Dr. Gerald Seider Company Profile Consulting- & Engineering Services Simulation and Analysis of complex fluid flow and
More informationDriven Cavity Example
BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square
More informationFast Dynamic Load Balancing for Extreme Scale Systems
Fast Dynamic Load Balancing for Extreme Scale Systems Cameron W. Smith, Gerrett Diamond, M.S. Shephard Computation Research Center (SCOREC) Rensselaer Polytechnic Institute Outline: n Some comments on
More informationLattice Boltzmann Method: A New Alternative for Loss Estimation Juan C. Díaz Cruz* a, Sergio E. Rodríguez a, Sergio A. Ordóñez b, Felipe Muñoz a a
799 VOL. 31, 2013 CHEMICAL ENGINEERING TRANSACTIONS A publication of The Italian Association of Chemical Engineering Online at: www.aidic.it/cet Guest Editors: Eddy De Rademaeker, Bruno Fabiano, Simberto
More informationPossibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics
Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Masanori Hashiguchi 1 1 Keisoku Engineering System Co., Ltd. 1-9-5 Uchikanda, Chiyoda-ku,
More informationAbout Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS.
About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS. Phoenix FD core SIMULATION & RENDERING. SIMULATION CORE - GRID-BASED
More informationAdvanced 1D/2D Modeling Using HEC-RAS
Advanced 1D/2D Modeling Using HEC-RAS Davis, California Objectives This is an advanced course in applying computer program HEC-RAS. The course provides participants with the knowledge to effectively use
More informationCGT 581 G Fluids. Overview. Some terms. Some terms
CGT 581 G Fluids Bedřich Beneš, Ph.D. Purdue University Department of Computer Graphics Technology Overview Some terms Incompressible Navier-Stokes Boundary conditions Lagrange vs. Euler Eulerian approaches
More informationEfficient Algorithmic Approaches for Flow Simulations on Cartesian Grids
Efficient Algorithmic Approaches for Flow Simulations on Cartesian Grids M. Bader, H.-J. Bungartz, B. Gatzhammer, M. Mehl, T. Neckel, T. Weinzierl TUM Department of Informatics Chair of Scientific Computing
More information