Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm

Size: px
Start display at page:

Download "Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm"

Transcription

1 Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm March 17 March 21, 2014 Florian Schornbaum, Martin Bauer, Simon Bogner Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany

2 Introductory Lecture Monday, March 17, 2014

3 Outline Introduction (of us / of participants) The walberla Framework What is the walberla Framework? Examples / Applications Course Schedule Introduction to the Lattice Boltzmann Method Introduction to the walberla Framework Data Structures / Underlying Concepts Domain Decomposition & Parallelization A Prototypical Simulation 3

4 Introduction (of us / of participants) Who are we? Who are you?

5 Introduction (of us / of participants) Martin Bauer [computational FAU] PhD LSS (chair for system simulation) core developer of walberla framework tutor for this compact course: Monday Friday Simon Bogner [math & computer FAU] PhD LSS (chair for system simulation) models and methods for lattice Boltzmann tutor for lectures on Tuesday & Wednesday Florian Schornbaum [computer FAU] PhD LSS (chair for system simulation) core developer of walberla framework tutor for this compact course: Monday Friday 5

6 Introduction (of us / of participants) Chair for System Simulation (1 st floor) Erlangen The University of Erlangen-Nuremberg [Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU)] has five faculties with over 270 chairs. 6

7 Introduction (of us / of participants) Chair for System Simulation (1 st floor) Faculty of Engineering (marked: Chair for System Simulation) FAU Erlangen-Nürnberg Faculty of Engineering Department of Computer Science Chair for System Simulation 7

8 Introduction (of us / of participants) Who are you?? Name? Field of study? Semester? Know-how in (computational) fluid mechanics? Programming Experience? (especially C++? C? Java?) 8

9 The walberla Framework What is the walberla Framework? Examples / Applications

10 The walberla Simulation Framework written in C++(11) main focus on CFD simulations based on the lattice Boltzmann method (LBM) but generally suitable for all kinds of numeric codes working with uniform domain decompositions (LSE solvers, phase field method) at its very core designed as an HPC software framework: scales from laptops to current petascale supercomputers largest simulation: 1,835,008 processes (IBM Blue Jülich) hybrid parallelization: MPI + OpenMP vectorization of compute kernels open source widely applicable lattice Boltzmann framework from Erlangen 10

11 The walberla Simulation Framework written in C++(11) main focus on CFD simulations based on the lattice Boltzmann method (LBM) but generally suitable for all kinds of numeric codes working with uniform domain decompositions (LSE solvers, phase field method) at its very core designed as an HPC software framework: scales from laptops to current petascale supercomputers largest simulation: 1,835,008 processes (IBM Blue Jülich) hybrid parallelization: MPI + OpenMP vectorization of compute kernels open source 11

12 The walberla Simulation Framework coupling with in-house rigid body physics engine pe automated build and test system: New git commit to central repository triggers builds on different systems with different compilers and different configurations (MPI on/off, OpenMP on/off, Debug/Release, float/double, ). After successful build, unit tests & set of applications are executed. support for different platforms (Linux, Windows) and compilers llvm/clang 12

13 The walberla Simulation Framework blood flow through coronary artery tree (C. Godenschwager) [ C. Godenschwager, F. Schornbaum, M. Bauer, H. Köstler, and U. Rüde, A Framework for Hybrid Parallel Flow Simulations with a Trillion Cells in Complex Geometries, SC13, Denver] domain decomposition of complex geometry 13

14 The walberla Simulation Framework electron beam melting metal powder: 3D printing (R. Ammer, M. Markl) [ liquid: walberla (LBM) metal powder: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 14

15 The walberla Simulation Framework electron beam melting metal powder: 3D printing (R. Ammer, M. Markl) [ liquid: walberla (LBM) metal powder: pe (fully resolved rigid bodies) ] camera footage of real application 15

16 The walberla Simulation Framework liquid-gas-solid flow simulation: stable floating positions of boxes with different densities (S. Bogner) [ liquid: walberla (LBM) boxes: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 16

17 The walberla Simulation Framework liquid-gas-solid flow simulation: bubble rising through pearls hovering under water (S. Bogner) [ liquid: walberla (LBM) pearls: pe (fully resolved rigid bodies) ] POV-Ray rendering of actual simulation 17

18 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) THVZs 18

19 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) DNS (direct numerical simulation) Reynolds number: 1000 / D3Q19 TRT 4300 SuperMUC 101,466,432 fluid cells 25,800 blocks with 16 x 16 x 16 cells 19

20 The walberla Simulation Framework study of the vocal fold LBM with grid refinement (F. Schornbaum, E. Fattahi) number of different grid levels: time steps / sec (finest grid) total number of time steps (finest grid): 864,000 without refinement: 55.2 times more memory and 98.6 times the workload 20

21 Course Schedule

22 Course Schedule Monday Tuesday Wednesday Thursday Friday Introduction [Lecture] LB [Lecture] LB [Lecture] HPC (MPI) [Lecture] Project Presentations C++ [Lecture] C++ [Lecture] LB [Lecture] Work on Projects Project Presentations Lunch Break Lunch Break Lunch Break Lunch Break Lunch Break Getting to Know the Environment [Hands-On] walberla Tutorials [Hands-On] Introduction of Projects Work on Projects Work on Projects Work on Projects Morning time slots: and Lunch Break: Afternoon: (Friday: ) 22

23 Introduction to the Lattice Boltzmann Method

24 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Macroscopic quantities (velocity, density, ) can be calculated from the particle distribution functions. 24

25 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Macroscopic quantities (velocity, density, ) can be calculated from the particle distribution functions. 25

26 Introduction to the LBM regular grid with multiple particle distribution functions per cell particle distribution function = scalar value different models: D2Q9, D3Q19, D3Q27, Every particle distribution function is associated with a certain direction. 26

27 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 27

28 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 28

29 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] 29

30 Introduction to the LBM explicit method time stepping (separated into two steps) two steps: stream & collide streaming: values are copied to neighboring cells collision: All values of each cell are updated using only the values of this cell. [cell-local operation!] For the collision, different operators exist: SRT, TRT, MRT 30

31 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. These calculated values depend on the underlying boundary condition type. 31

32 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 32

33 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 33

34 Introduction to the LBM boundary treatment pre-streaming step boundary Particle distribution functions are calculated for boundary cells which are neighboring fluid cells. During streaming, these values are transferred/copied into the neighboring fluid cells. 34

35 Introduction to the LBM boundary treatment pre-streaming step boundary The lattice Boltzmann method allows to handle very complex geometries easily! Cells are typically just marked as being a boundary cell of a certain type: no-slip, free slip, velocity bounce-back, pressure, open boundary, [ flag field = special grid that stores these markings] 35

36 Introduction to the LBM A typical time step of the LBM looks like as follows: 1. boundary treatment is performed (= values in the boundary cells are prepared for streaming) 2. the streaming step is executed (requires nearest neighbor access) 3. the collision operator is evaluated (cell-local operation) A typical parallel time step includes communication: 1. all values of cells on the border of a process s partition of the domain are exchanged between neighboring processes (involves so-called ghost layer cells) 2. boundary treatment is performed (= values in the boundary cells are prepared for streaming) 3. the streaming step is executed (requires nearest neighbor access) 4. the collision operator is evaluated (cell-local operation) 36

37 Introduction to walberla Data Structures / Underlying Concepts Domain Decomposition & Parallelization A Prototypical Simulation

38 Introduction to walberla geometry given by surface mesh domain decomposition into blocks flow simulation only in here (example: complex geometry of an artery) load balancing empty blocks are discarded Load balancing can be based on either space-filling curves (Z-order/Morton order, Hilbert curve) using the underlying forest of octrees or graph partitioning (METIS, ). Whatever fits best the needs of the simulation. 38

39 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded Load balancing can be based on either space-filling curves (Z-order/Morton order, Hilbert curve) using the underlying forest of octrees or graph partitioning (METIS, ). Whatever fits best the needs of the simulation. 39

40 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded allocation of block data ( grids) The domain decomposition and load balancing can be performed during the actual simulation OR 40

41 Introduction to walberla geometry given by surface mesh domain decomposition into blocks load balancing empty blocks are discarded separation of domain partitioning from simulation DISK DISK file size: kilobytes to few megabytes allocation of block data ( grids) 41

42 Introduction to walberla geometry given by surface mesh domain decomposition into blocks All of this (the entire pipeline) works just the same when grid refinement is used. load balancing empty blocks are discarded separation of domain partitioning from simulation DISK DISK file size: kilobytes to few megabytes allocation of block data ( grids) 42

43 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids grid refinement: octree-like decomposition special case of the much more general forest of octrees data structure non-uniform/refined grids forest of octrees: each block contains a uniform grid of the same size 2:1 balance between neighboring cells on level transitions (enables LB refinement algorithms) 43

44 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids special case of the much more general forest of octrees data structure non-uniform/refined grids octree-like decomposition of a globally uniform grid forest of octrees: each block contains a uniform grid with cells of the same size performance of uniform grids & good fitting to complex geometries 44

45 Introduction to walberla Domain Decomposition: regular decomposition into blocks containing uniform grids octree-like decomposition of the simulation space?????????????????????????????? special case of the much more general forest of octrees data structure non-uniform/refined grids forest of octrees: In general, each block contains whatever data was registered by the application: can be any arbitrary (sub)grid, scalar values, any object (in terms of C++), 45

46 Introduction to walberla Parallelization : data exchange on borders between blocks via ghost layers sender process receiver process (slightly more complicated for non-uniform domain decompositions, but the same general ideas still apply) typical parallelization of the underlying algorithm (e.g., LB): 1. perform communication: copy outermost layer of inner cells of perform communication: sender to ghost layer of receiver 2. perform one time step of the computation algorithm only on all inner cells of each block this computation algorithm is only allowed to access nearest neighbor cells during calculations! 46

47 Introduction to walberla A Prototypical Simulation: 1. construct domain decomposition (includes load balancing = distribution of blocks to all available processes) 2. add data to blocks (LB: at least one grid that stores the particle distribution functions and one grid that is used for marking cells as either being fluid or obstacle additional grids/classes may be required/useful [e.g., an object that is responsible for the boundary treatment]) 3. set-up geometry/boundaries (= mark cells as either being fluid or a boundary of a certain type information about the geometry/boundaries can be loaded from a mesh file, read in from a configuration file, set in the source code, etc.) 4. specify the algorithms/functions that are supposed to be executed in each time step (basic LB: communication, boundary treatment, stream & collide, output[optional]) 5. run the simulation = execute fixed number of time steps 47

48 THANK YOU FOR YOUR ATTENTION! QUESTIONS? (visit

simulation framework for piecewise regular grids

simulation framework for piecewise regular grids WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler

More information

Peta-Scale Simulations with the HPC Software Framework walberla:

Peta-Scale Simulations with the HPC Software Framework walberla: Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,

More information

walberla: Developing a Massively Parallel HPC Framework

walberla: Developing a Massively Parallel HPC Framework walberla: Developing a Massively Parallel HPC Framework SIAM CS&E 2013, Boston February 26, 2013 Florian Schornbaum*, Christian Godenschwager*, Martin Bauer*, Matthias Markl, Ulrich Rüde* *Chair for System

More information

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla SIAM PP 2016, April 13 th 2016 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer,

More information

Massively Parallel Phase Field Simulations using HPC Framework walberla

Massively Parallel Phase Field Simulations using HPC Framework walberla Massively Parallel Phase Field Simulations using HPC Framework walberla SIAM CSE 2015, March 15 th 2015 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer, Harald Köstler and Ulrich

More information

The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms

The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms Harald Köstler, Uli Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

A Python extension for the massively parallel framework walberla

A Python extension for the massively parallel framework walberla A Python extension for the massively parallel framework walberla PyHPC at SC 14, November 17 th 2014 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler

More information

(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more

(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more Parallel Free-Surface Extension of the Lattice-Boltzmann Method A Lattice-Boltzmann Approach for Simulation of Two-Phase Flows Stefan Donath (LSS Erlangen, stefan.donath@informatik.uni-erlangen.de) Simon

More information

Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method

Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method June 21, 2011 Introduction Free Surface LBM Liquid-Gas-Solid Flows Parallel Computing Examples and More References Fig. Simulation

More information

Lattice Boltzmann Methods on the way to exascale

Lattice Boltzmann Methods on the way to exascale Lattice Boltzmann Methods on the way to exascale Ulrich Rüde (LSS Erlangen, ulrich.ruede@fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de HIGH PERFORMANCE

More information

Software and Performance Engineering for numerical codes on GPU clusters

Software and Performance Engineering for numerical codes on GPU clusters Software and Performance Engineering for numerical codes on GPU clusters H. Köstler International Workshop of GPU Solutions to Multiscale Problems in Science and Engineering Harbin, China 28.7.2010 2 3

More information

Simulation of moving Particles in 3D with the Lattice Boltzmann Method

Simulation of moving Particles in 3D with the Lattice Boltzmann Method Simulation of moving Particles in 3D with the Lattice Boltzmann Method, Nils Thürey, Christian Feichtinger, Hans-Joachim Schmid Chair for System Simulation University Erlangen/Nuremberg Chair for Particle

More information

Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures

Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Jan Treibig, Simon Hausmann, Ulrich Ruede Zusammenfassung The Lattice Boltzmann method (LBM) is a well established algorithm

More information

Lattice Boltzmann with CUDA

Lattice Boltzmann with CUDA Lattice Boltzmann with CUDA Lan Shi, Li Yi & Liyuan Zhang Hauptseminar: Multicore Architectures and Programming Page 1 Outline Overview of LBM An usage of LBM Algorithm Implementation in CUDA and Optimization

More information

A Contact Angle Model for the Parallel Free Surface Lattice Boltzmann Method in walberla Stefan Donath (stefan.donath@informatik.uni-erlangen.de) Computer Science 10 (System Simulation) University of Erlangen-Nuremberg

More information

International Supercomputing Conference 2009

International Supercomputing Conference 2009 International Supercomputing Conference 2009 Implementation of a Lattice-Boltzmann-Method for Numerical Fluid Mechanics Using the nvidia CUDA Technology E. Riegel, T. Indinger, N.A. Adams Technische Universität

More information

LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS

LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS 14 th European Conference on Mixing Warszawa, 10-13 September 2012 LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS Felix Muggli a, Laurent Chatagny a, Jonas Lätt b a Sulzer Markets & Technology

More information

Reconstruction of Trees from Laser Scan Data and further Simulation Topics

Reconstruction of Trees from Laser Scan Data and further Simulation Topics Reconstruction of Trees from Laser Scan Data and further Simulation Topics Helmholtz-Research Center, Munich Daniel Ritter http://www10.informatik.uni-erlangen.de Overview 1. Introduction of the Chair

More information

High Performance Computing

High Performance Computing High Performance Computing ADVANCED SCIENTIFIC COMPUTING Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich

More information

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation) FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) walberla: Visualization of Fluid

More information

From Notebooks to Supercomputers: Tap the Full Potential of Your CUDA Resources with LibGeoDecomp

From Notebooks to Supercomputers: Tap the Full Potential of Your CUDA Resources with LibGeoDecomp From Notebooks to Supercomputers: Tap the Full Potential of Your CUDA Resources with andreas.schaefer@cs.fau.de Friedrich-Alexander-Universität Erlangen-Nürnberg GPU Technology Conference 2013, San José,

More information

Navier-Stokes & Flow Simulation

Navier-Stokes & Flow Simulation Last Time? Navier-Stokes & Flow Simulation Pop Worksheet! Teams of 2. Hand in to Jeramey after we discuss. Sketch the first few frames of a 2D explicit Euler mass-spring simulation for a 2x3 cloth network

More information

Performance and Software-Engineering Considerations for Massively Parallel Simulations

Performance and Software-Engineering Considerations for Massively Parallel Simulations Performance and Software-Engineering Considerations for Massively Parallel Simulations Ulrich Rüde (ruede@cs.fau.de) Ben Bergen, Frank Hülsemann, Christoph Freundl Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

Numerical Algorithms on Multi-GPU Architectures

Numerical Algorithms on Multi-GPU Architectures Numerical Algorithms on Multi-GPU Architectures Dr.-Ing. Harald Köstler 2 nd International Workshops on Advances in Computational Mechanics Yokohama, Japan 30.3.2010 2 3 Contents Motivation: Applications

More information

Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method

Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method G. Wellein, T. Zeiser, G. Hager HPC Services Regional Computing Center A. Nitsure, K. Iglberger, U. Rüde Chair for System

More information

Parallel Simulation of Dendritic Growth On Unstructured Grids

Parallel Simulation of Dendritic Growth On Unstructured Grids Parallel Simulation of Dendritic Growth On Unstructured Grids, Julian Hammer, Dietmar Fey Friedrich-Alexander-Universität Erlangen-Nürnberg IA 3 Workshop @SC11, Nov. 13th, 2011 Outline 1 What and why?

More information

Technical Report TR

Technical Report TR Technical Report TR-2015-09 Boundary condition enforcing methods for smoothed particle hydrodynamics Arman Pazouki 1, Baofang Song 2, Dan Negrut 1 1 University of Wisconsin-Madison, Madison, WI, 53706-1572,

More information

Lattice Boltzmann methods on the way to exascale

Lattice Boltzmann methods on the way to exascale Lattice Boltzmann methods on the way to exascale Ulrich Rüde LSS Erlangen and CERFACS Toulouse ulrich.ruede@fau.de Centre Européen de Recherche et de Formation Avancée en Calcul Scientifique www.cerfacs.fr

More information

Navier-Stokes & Flow Simulation

Navier-Stokes & Flow Simulation Last Time? Navier-Stokes & Flow Simulation Implicit Surfaces Marching Cubes/Tetras Collision Detection & Response Conservative Bounding Regions backtracking fixing Today Flow Simulations in Graphics Flow

More information

CS GPU and GPGPU Programming Lecture 2: Introduction; GPU Architecture 1. Markus Hadwiger, KAUST

CS GPU and GPGPU Programming Lecture 2: Introduction; GPU Architecture 1. Markus Hadwiger, KAUST CS 380 - GPU and GPGPU Programming Lecture 2: Introduction; GPU Architecture 1 Markus Hadwiger, KAUST Reading Assignment #2 (until Feb. 17) Read (required): GLSL book, chapter 4 (The OpenGL Programmable

More information

LARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS

LARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS ECCOMAS Congress 2016 VII European Congress on Computational Methods in Applied Sciences and Engineering M. Papadrakakis, V. Papadopoulos, G. Stefanou, V. Plevris (eds.) Crete Island, Greece, 5 10 June

More information

Towards PetaScale Computational Science

Towards PetaScale Computational Science Towards PetaScale Computational Science U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with many Lehrstuhl für Informatik 10 (Systemsimulation) Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

Navier-Stokes & Flow Simulation

Navier-Stokes & Flow Simulation Last Time? Navier-Stokes & Flow Simulation Optional Reading for Last Time: Spring-Mass Systems Numerical Integration (Euler, Midpoint, Runge-Kutta) Modeling string, hair, & cloth HW2: Cloth & Fluid Simulation

More information

Large scale Imaging on Current Many- Core Platforms

Large scale Imaging on Current Many- Core Platforms Large scale Imaging on Current Many- Core Platforms SIAM Conf. on Imaging Science 2012 May 20, 2012 Dr. Harald Köstler Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen,

More information

Domain Decomposition for Colloid Clusters. Pedro Fernando Gómez Fernández

Domain Decomposition for Colloid Clusters. Pedro Fernando Gómez Fernández Domain Decomposition for Colloid Clusters Pedro Fernando Gómez Fernández MSc in High Performance Computing The University of Edinburgh Year of Presentation: 2004 Authorship declaration I, Pedro Fernando

More information

Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids

Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Sebastian Kuckuk FAU Erlangen-Nürnberg 18.01.2016 HiStencils 2016, Prague, Czech Republic Outline Outline Scope

More information

CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman)

CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC. Guest Lecturer: Sukhyun Song (original slides by Alan Sussman) CMSC 714 Lecture 6 MPI vs. OpenMP and OpenACC Guest Lecturer: Sukhyun Song (original slides by Alan Sussman) Parallel Programming with Message Passing and Directives 2 MPI + OpenMP Some applications can

More information

CAF versus MPI Applicability of Coarray Fortran to a Flow Solver

CAF versus MPI Applicability of Coarray Fortran to a Flow Solver CAF versus MPI Applicability of Coarray Fortran to a Flow Solver Manuel Hasert, Harald Klimach, Sabine Roller m.hasert@grs-sim.de Applied Supercomputing in Engineering Motivation We develop several CFD

More information

Joint Advanced Student School 2007 Martin Dummer

Joint Advanced Student School 2007 Martin Dummer Sierpiński-Curves Joint Advanced Student School 2007 Martin Dummer Statement of the Problem What is the best way to store a triangle mesh efficiently in memory? The following points are desired : Easy

More information

A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla

A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler, and Ulrich

More information

SPEED-UP GEARBOX SIMULATIONS BY INTEGRATING SCORG. Dr. Christine Klier, Sahand Saheb-Jahromi, Ludwig Berger*

SPEED-UP GEARBOX SIMULATIONS BY INTEGRATING SCORG. Dr. Christine Klier, Sahand Saheb-Jahromi, Ludwig Berger* SPEED-UP GEARBOX SIMULATIONS BY INTEGRATING SCORG Dr. Christine Klier, Sahand Saheb-Jahromi, Ludwig Berger* CFD SCHUCK ENGINEERING Engineering Services in computational fluid Dynamics (CFD) 25 employees

More information

High-Performance Scientific Computing

High-Performance Scientific Computing High-Performance Scientific Computing Instructor: Randy LeVeque TA: Grady Lemoine Applied Mathematics 483/583, Spring 2011 http://www.amath.washington.edu/~rjl/am583 World s fastest computers http://top500.org

More information

Massively Parallel Finite Element Simulations with deal.ii

Massively Parallel Finite Element Simulations with deal.ii Massively Parallel Finite Element Simulations with deal.ii Timo Heister, Texas A&M University 2012-02-16 SIAM PP2012 joint work with: Wolfgang Bangerth, Carsten Burstedde, Thomas Geenen, Martin Kronbichler

More information

Two-Phase flows on massively parallel multi-gpu clusters

Two-Phase flows on massively parallel multi-gpu clusters Two-Phase flows on massively parallel multi-gpu clusters Peter Zaspel Michael Griebel Institute for Numerical Simulation Rheinische Friedrich-Wilhelms-Universität Bonn Workshop Programming of Heterogeneous

More information

Sustainability and Efficiency for Simulation Software in the Exascale Era

Sustainability and Efficiency for Simulation Software in the Exascale Era Sustainability and Efficiency for Simulation Software in the Exascale Era Dominik Thönnes, Ulrich Rüde, Nils Kohl Chair for System Simulation, University of Erlangen-Nürnberg March 09, 2018 SIAM Conference

More information

Free Surface Lattice-Boltzmann fluid simulations. with and without level sets.

Free Surface Lattice-Boltzmann fluid simulations. with and without level sets. ree Surface Lattice-Boltzmann fluid simulations with and without level sets Nils Thürey, Ulrich Rüde University of Erlangen-Nuremberg System Simulation roup Cauerstr. 6, 91054 Erlangen, ermany Email: Nils.Thuerey@cs.fau.de

More information

Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures

Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Dirk Ribbrock, Markus Geveler, Dominik Göddeke, Stefan Turek Angewandte Mathematik, Technische Universität Dortmund

More information

Webinar #3 Lattice Boltzmann method for CompBioMed (incl. Palabos)

Webinar #3 Lattice Boltzmann method for CompBioMed (incl. Palabos) Webinar series A Centre of Excellence in Computational Biomedicine Webinar #3 Lattice Boltzmann method for CompBioMed (incl. Palabos) 19 March 2018 The webinar will start at 12pm CET / 11am GMT Dr Jonas

More information

HIGH PERFORMANCE COMPUTATION (HPC) FOR THE

HIGH PERFORMANCE COMPUTATION (HPC) FOR THE HIGH PERFORMANCE COMPUTATION (HPC) FOR THE DEVELOPMENT OF FLUIDIZED BED TECHNOLOGIES FOR BIOMASS GASIFICATION AND CO2 CAPTURE P. Fede, H. Neau, O. Simonin Université de Toulouse; INPT, UPS ; IMFT ; 31400

More information

Architecture Aware Multigrid

Architecture Aware Multigrid Architecture Aware Multigrid U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with D. Ritter, T. Gradl, M. Stürmer, H. Köstler, J. Treibig and many more students Lehrstuhl für Informatik 10 (Systemsimulation)

More information

Visualization. Images are used to aid in understanding of data. Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [chapter 26]

Visualization. Images are used to aid in understanding of data. Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [chapter 26] Visualization Images are used to aid in understanding of data Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [chapter 26] Tumor SCI, Utah Scientific Visualization Visualize large

More information

Introduction CPS343. Spring Parallel and High Performance Computing. CPS343 (Parallel and HPC) Introduction Spring / 29

Introduction CPS343. Spring Parallel and High Performance Computing. CPS343 (Parallel and HPC) Introduction Spring / 29 Introduction CPS343 Parallel and High Performance Computing Spring 2018 CPS343 (Parallel and HPC) Introduction Spring 2018 1 / 29 Outline 1 Preface Course Details Course Requirements 2 Background Definitions

More information

CAF versus MPI - Applicability of Coarray Fortran to a Flow Solver

CAF versus MPI - Applicability of Coarray Fortran to a Flow Solver CAF versus MPI - Applicability of Coarray Fortran to a Flow Solver Manuel Hasert, Harald Klimach, Sabine Roller 2 German Research School for Simulation Sciences GmbH, 52062 Aachen, Germany 2 RWTH Aachen

More information

Dynamic Mode Decomposition analysis of flow fields from CFD Simulations

Dynamic Mode Decomposition analysis of flow fields from CFD Simulations Dynamic Mode Decomposition analysis of flow fields from CFD Simulations Technische Universität München Thomas Indinger Lukas Haag, Daiki Matsumoto, Christoph Niedermeier in collaboration with Agenda Motivation

More information

SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS

SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22, 2012 SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC

More information

Tutorial 17. Using the Mixture and Eulerian Multiphase Models

Tutorial 17. Using the Mixture and Eulerian Multiphase Models Tutorial 17. Using the Mixture and Eulerian Multiphase Models Introduction: This tutorial examines the flow of water and air in a tee junction. First you will solve the problem using the less computationally-intensive

More information

BOUNDLESS FLUIDS USING THE LATTICE-BOLTZMANN METHOD. A Thesis. Presented to. the Faculty of California Polytechnic State University.

BOUNDLESS FLUIDS USING THE LATTICE-BOLTZMANN METHOD. A Thesis. Presented to. the Faculty of California Polytechnic State University. BOUNDLESS FLUIDS USING THE LATTICE-BOLTZMANN METHOD A Thesis Presented to the Faculty of California Polytechnic State University San Luis Obispo In Partial Fulfillment of the Requirements for the Degree

More information

Forest-of-octrees AMR: algorithms and interfaces

Forest-of-octrees AMR: algorithms and interfaces Forest-of-octrees AMR: algorithms and interfaces Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac, Georg Stadler, Lucas C. Wilcox Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

Express Introductory Training in ANSYS Fluent Workshop 02 Using the Discrete Phase Model (DPM)

Express Introductory Training in ANSYS Fluent Workshop 02 Using the Discrete Phase Model (DPM) Express Introductory Training in ANSYS Fluent Workshop 02 Using the Discrete Phase Model (DPM) Dimitrios Sofialidis Technical Manager, SimTec Ltd. Mechanical Engineer, PhD PRACE Autumn School 2013 - Industry

More information

NVIDIA. Interacting with Particle Simulation in Maya using CUDA & Maximus. Wil Braithwaite NVIDIA Applied Engineering Digital Film

NVIDIA. Interacting with Particle Simulation in Maya using CUDA & Maximus. Wil Braithwaite NVIDIA Applied Engineering Digital Film NVIDIA Interacting with Particle Simulation in Maya using CUDA & Maximus Wil Braithwaite NVIDIA Applied Engineering Digital Film Some particle milestones FX Rendering Physics 1982 - First CG particle FX

More information

CUDA. Fluid simulation Lattice Boltzmann Models Cellular Automata

CUDA. Fluid simulation Lattice Boltzmann Models Cellular Automata CUDA Fluid simulation Lattice Boltzmann Models Cellular Automata Please excuse my layout of slides for the remaining part of the talk! Fluid Simulation Navier Stokes equations for incompressible fluids

More information

GPU-based Distributed Behavior Models with CUDA

GPU-based Distributed Behavior Models with CUDA GPU-based Distributed Behavior Models with CUDA Courtesy: YouTube, ISIS Lab, Universita degli Studi di Salerno Bradly Alicea Introduction Flocking: Reynolds boids algorithm. * models simple local behaviors

More information

Realtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research

Realtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research 1 Realtime Water Simulation on GPU Nuttapong Chentanez NVIDIA Research 2 3 Overview Approaches to realtime water simulation Hybrid shallow water solver + particles Hybrid 3D tall cell water solver + particles

More information

Asynchronous OpenCL/MPI numerical simulations of conservation laws

Asynchronous OpenCL/MPI numerical simulations of conservation laws Asynchronous OpenCL/MPI numerical simulations of conservation laws Philippe HELLUY 1,3, Thomas STRUB 2. 1 IRMA, Université de Strasbourg, 2 AxesSim, 3 Inria Tonus, France IWOCL 2015, Stanford Conservation

More information

Supersonic Flow Over a Wedge

Supersonic Flow Over a Wedge SPC 407 Supersonic & Hypersonic Fluid Dynamics Ansys Fluent Tutorial 2 Supersonic Flow Over a Wedge Ahmed M Nagib Elmekawy, PhD, P.E. Problem Specification A uniform supersonic stream encounters a wedge

More information

High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters

High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters SIAM PP 2014 High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters C. Riesinger, A. Bakhtiari, M. Schreiber Technische Universität München February 20, 2014

More information

Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 12] April 23, 2002 Frank Pfenning Carnegie Mellon University

Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 12] April 23, 2002 Frank Pfenning Carnegie Mellon University 15-462 Computer Graphics I Lecture 21 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 12] April 23, 2002 Frank Pfenning Carnegie Mellon University http://www.cs.cmu.edu/~fp/courses/graphics/

More information

Introduction to the immersed boundary method

Introduction to the immersed boundary method Introduction to the immersed boundary method Motivation. Hydrodynamics and boundary conditions The incompressible Navier-Stokes equations, ( ) u ρ + (u )u = p + ρν 2 u + f, () t are partial differential

More information

Adaptive Refinement Tree (ART) code. N-Body: Parallelization using OpenMP and MPI

Adaptive Refinement Tree (ART) code. N-Body: Parallelization using OpenMP and MPI Adaptive Refinement Tree (ART) code N-Body: Parallelization using OpenMP and MPI 1 Family of codes N-body: OpenMp N-body: MPI+OpenMP N-body+hydro+cooling+SF: OpenMP N-body+hydro+cooling+SF: MPI 2 History:

More information

PHYSICALLY BASED ANIMATION

PHYSICALLY BASED ANIMATION PHYSICALLY BASED ANIMATION CS148 Introduction to Computer Graphics and Imaging David Hyde August 2 nd, 2016 WHAT IS PHYSICS? the study of everything? WHAT IS COMPUTATION? the study of everything? OUTLINE

More information

LATTICE-BOLTZMANN AND COMPUTATIONAL FLUID DYNAMICS

LATTICE-BOLTZMANN AND COMPUTATIONAL FLUID DYNAMICS LATTICE-BOLTZMANN AND COMPUTATIONAL FLUID DYNAMICS NAVIER-STOKES EQUATIONS u t + u u + 1 ρ p = Ԧg + ν u u=0 WHAT IS COMPUTATIONAL FLUID DYNAMICS? Branch of Fluid Dynamics which uses computer power to approximate

More information

Free Surface Flows with Moving and Deforming Objects for LBM

Free Surface Flows with Moving and Deforming Objects for LBM Free Surface Flows with Moving and Deforming Objects for LBM Nils Thürey, Klaus Iglberger, Ulrich Rüde Institute for System Simulation (LSS), University of Erlangen-Nuremberg, Cauerstr. 6, 91058 Erlangen,

More information

Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs)

Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) OBJECTIVE FLUID SIMULATIONS Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) The basic objective of the project is the implementation of the paper Stable Fluids (Jos Stam, SIGGRAPH 99). The final

More information

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2

More information

NIA CFD Futures Conference Hampton, VA; August 2012

NIA CFD Futures Conference Hampton, VA; August 2012 Petascale Computing and Similarity Scaling in Turbulence P. K. Yeung Schools of AE, CSE, ME Georgia Tech pk.yeung@ae.gatech.edu NIA CFD Futures Conference Hampton, VA; August 2012 10 2 10 1 10 4 10 5 Supported

More information

Shape of Things to Come: Next-Gen Physics Deep Dive

Shape of Things to Come: Next-Gen Physics Deep Dive Shape of Things to Come: Next-Gen Physics Deep Dive Jean Pierre Bordes NVIDIA Corporation Free PhysX on CUDA PhysX by NVIDIA since March 2008 PhysX on CUDA available: August 2008 GPU PhysX in Games Physical

More information

Directions: 1) Delete this text box 2) Insert desired picture here

Directions: 1) Delete this text box 2) Insert desired picture here Directions: 1) Delete this text box 2) Insert desired picture here Multi-Disciplinary Applications using Overset Grid Technology in STAR-CCM+ CD-adapco Dmitry Pinaev, Frank Schäfer, Eberhard Schreck Outline

More information

Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede

Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Preliminary Experiences with the Uintah Framework on on Intel Xeon Phi and Stampede Qingyu Meng, Alan Humphrey, John Schmidt, Martin Berzins Thanks to: TACC Team for early access to Stampede J. Davison

More information

Technical Tricks of Coarse-Grained MD Visualization with VMD

Technical Tricks of Coarse-Grained MD Visualization with VMD Technical Tricks of Coarse-Grained MD Visualization with VMD Institut für Computerphysik, Universität Stuttgart Stuttgart, Germany Arbitrary Units Lennard-Jones-Potential All-atom MD simulations Lengths

More information

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications

A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications A Scalable Adaptive Mesh Refinement Framework For Parallel Astrophysics Applications James Bordner, Michael L. Norman San Diego Supercomputer Center University of California, San Diego 15th SIAM Conference

More information

Generic finite element capabilities for forest-of-octrees AMR

Generic finite element capabilities for forest-of-octrees AMR Generic finite element capabilities for forest-of-octrees AMR Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

11/1/13. Visualization. Scientific Visualization. Types of Data. Height Field. Contour Curves. Meshes

11/1/13. Visualization. Scientific Visualization. Types of Data. Height Field. Contour Curves. Meshes CSCI 420 Computer Graphics Lecture 26 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 2.11] Jernej Barbic University of Southern California Scientific Visualization

More information

Visualization. CSCI 420 Computer Graphics Lecture 26

Visualization. CSCI 420 Computer Graphics Lecture 26 CSCI 420 Computer Graphics Lecture 26 Visualization Height Fields and Contours Scalar Fields Volume Rendering Vector Fields [Angel Ch. 11] Jernej Barbic University of Southern California 1 Scientific Visualization

More information

Optimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH

Optimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH Optimising MPI Applications for Heterogeneous Coupled Clusters with MetaMPICH Carsten Clauss, Martin Pöppe, Thomas Bemmerl carsten@lfbs.rwth-aachen.de http://www.mp-mpich.de Lehrstuhl für Betriebssysteme

More information

RAPID LARGE-SCALE CARTESIAN MESHING FOR AERODYNAMIC COMPUTATIONS

RAPID LARGE-SCALE CARTESIAN MESHING FOR AERODYNAMIC COMPUTATIONS RAPID LARGE-SCALE CARTESIAN MESHING FOR AERODYNAMIC COMPUTATIONS Daisuke Sasaki*, Kazuhiro Nakahashi** *Department of Aeronautics, Kanazawa Institute of Technology, **JAXA Keywords: Meshing, Cartesian

More information

Accurate and Efficient Turbomachinery Simulation. Chad Custer, PhD Turbomachinery Technical Specialist

Accurate and Efficient Turbomachinery Simulation. Chad Custer, PhD Turbomachinery Technical Specialist Accurate and Efficient Turbomachinery Simulation Chad Custer, PhD Turbomachinery Technical Specialist Outline Turbomachinery simulation advantages Axial fan optimization Description of design objectives

More information

Advanced Parallel Programming

Advanced Parallel Programming Sebastian von Alfthan Jussi Enkovaara Pekka Manninen Advanced Parallel Programming February 15-17, 2016 PRACE Advanced Training Center CSC IT Center for Science Ltd, Finland All material (C) 2011-2016

More information

MASSIVELY-PARALLEL MULTI-GPU SIMULATIONS FOR FAST AND ACCURATE AUTOMOTIVE AERODYNAMICS

MASSIVELY-PARALLEL MULTI-GPU SIMULATIONS FOR FAST AND ACCURATE AUTOMOTIVE AERODYNAMICS 6th European Conference on Computational Mechanics (ECCM 6) 7th European Conference on Computational Fluid Dynamics (ECFD 7) -5 June 28, Glasgow, UK MASSIVELY-PARALLEL MULTI-GPU SIMULATIONS FOR FAST AND

More information

CS 231. Fluid simulation

CS 231. Fluid simulation CS 231 Fluid simulation Why Simulate Fluids? Feature film special effects Computer games Medicine (e.g. blood flow in heart) Because it s fun Fluid Simulation Called Computational Fluid Dynamics (CFD)

More information

Parallel Mesh Multiplication for Code_Saturne

Parallel Mesh Multiplication for Code_Saturne Parallel Mesh Multiplication for Code_Saturne Pavla Kabelikova, Ales Ronovsky, Vit Vondrak a Dept. of Applied Mathematics, VSB-Technical University of Ostrava, Tr. 17. listopadu 15, 708 00 Ostrava, Czech

More information

Alex Li 11/20/2009. Chris Wojtan, Nils Thurey, Markus Gross, Greg Turk

Alex Li 11/20/2009. Chris Wojtan, Nils Thurey, Markus Gross, Greg Turk Alex Li 11/20/2009 Chris Wojtan, Nils Thurey, Markus Gross, Greg Turk duction Overview of Lagrangian of Topological s Altering the Topology 2 Presents a method for accurately tracking the moving surface

More information

STAR Global Conference 2012 Noordwijk, March 20-21, The InDesA Virtual Test Bench. Dr. Fabiano Bet Dr. Gerald Seider

STAR Global Conference 2012 Noordwijk, March 20-21, The InDesA Virtual Test Bench. Dr. Fabiano Bet Dr. Gerald Seider STAR Global Conference 2012 Noordwijk, March 20-21, 2012 The InDesA Dr. Fabiano Bet Dr. Gerald Seider Company Profile Consulting- & Engineering Services Simulation and Analysis of complex fluid flow and

More information

Driven Cavity Example

Driven Cavity Example BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square

More information

Fast Dynamic Load Balancing for Extreme Scale Systems

Fast Dynamic Load Balancing for Extreme Scale Systems Fast Dynamic Load Balancing for Extreme Scale Systems Cameron W. Smith, Gerrett Diamond, M.S. Shephard Computation Research Center (SCOREC) Rensselaer Polytechnic Institute Outline: n Some comments on

More information

Lattice Boltzmann Method: A New Alternative for Loss Estimation Juan C. Díaz Cruz* a, Sergio E. Rodríguez a, Sergio A. Ordóñez b, Felipe Muñoz a a

Lattice Boltzmann Method: A New Alternative for Loss Estimation Juan C. Díaz Cruz* a, Sergio E. Rodríguez a, Sergio A. Ordóñez b, Felipe Muñoz a a 799 VOL. 31, 2013 CHEMICAL ENGINEERING TRANSACTIONS A publication of The Italian Association of Chemical Engineering Online at: www.aidic.it/cet Guest Editors: Eddy De Rademaeker, Bruno Fabiano, Simberto

More information

Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics

Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Masanori Hashiguchi 1 1 Keisoku Engineering System Co., Ltd. 1-9-5 Uchikanda, Chiyoda-ku,

More information

About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS.

About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS. About Phoenix FD PLUGIN FOR 3DS MAX AND MAYA. SIMULATING AND RENDERING BOTH LIQUIDS AND FIRE/SMOKE. USED IN MOVIES, GAMES AND COMMERCIALS. Phoenix FD core SIMULATION & RENDERING. SIMULATION CORE - GRID-BASED

More information

Advanced 1D/2D Modeling Using HEC-RAS

Advanced 1D/2D Modeling Using HEC-RAS Advanced 1D/2D Modeling Using HEC-RAS Davis, California Objectives This is an advanced course in applying computer program HEC-RAS. The course provides participants with the knowledge to effectively use

More information

CGT 581 G Fluids. Overview. Some terms. Some terms

CGT 581 G Fluids. Overview. Some terms. Some terms CGT 581 G Fluids Bedřich Beneš, Ph.D. Purdue University Department of Computer Graphics Technology Overview Some terms Incompressible Navier-Stokes Boundary conditions Lagrange vs. Euler Eulerian approaches

More information

Efficient Algorithmic Approaches for Flow Simulations on Cartesian Grids

Efficient Algorithmic Approaches for Flow Simulations on Cartesian Grids Efficient Algorithmic Approaches for Flow Simulations on Cartesian Grids M. Bader, H.-J. Bungartz, B. Gatzhammer, M. Mehl, T. Neckel, T. Weinzierl TUM Department of Informatics Chair of Scientific Computing

More information