Lattice Boltzmann methods on the way to exascale

Size: px
Start display at page:

Download "Lattice Boltzmann methods on the way to exascale"

Transcription

1 Lattice Boltzmann methods on the way to exascale Ulrich Rüde LSS Erlangen and CERFACS Toulouse Centre Européen de Recherche et de Formation Avancée en Calcul Scientifique Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de 1

2 Outline Motivation Building Blocks for the Direct Simulation of Complex Flows 1. Supercomputing: scalable algorithms, efficient software 2. Solid phase - rigid body dynamics 3. Fluid phase - Lattice Boltzmann method 4. Electrostatics - finite volume 5. Fast implicit solvers - multigrid 6. Gas phase - free surface tracking, volume of fluids Multi-physics applications Coupling Examples Perspectives 2

3 Motivation Simulating Additive Manufacturing Bikas, H., Stavropoulos, P., & Chryssolouris, G. (2015). Additive manufacturing methods and modelling approaches: a critical review. The International Journal of Advanced Manufacturing Technology, Klassen, A., Scharowsky, T., & Körner, C. (2014). Evaporation model for beam based additive manufacturing using free surface lattice Boltzmann methods. Journal of Physics D: Applied Physics, 47(27),

4 Electron Beam Melting Process 3D printing EU-Project Fast- EBM ARCAM (Gothenburg) TWI (Cambridge) FAU Erlangen Generation of powder bed Energy transfer by electron beam penetration depth heat transfer Flow dynamics melting melt flow surface tension wetting capillary forces contact angles solidification Ammer, R., Markl, M., Ljungblad, U., Körner, C., & UR (2014). Simulating fast electron beam melting with a parallel thermal free surface lattice Boltzmann method. Computers & Mathematics with Applications, 67(2), Ammer, R., UR, Markl, M., Jüchter V., & Körner, C. (2014). Validation experiments for LBM simulations of electron beam melting. International Journal of Modern Physics C. 4

5 SuperMuc: 3 PFlops Building Block I: Current and Future High Performance Supercomputers 5

6 Multi-PetaFlops Supercomputers Sunway TaihuLight JUQUEEN SuperMUC (phase 1) SW26010 processor 10,649,600 cores 260 cores (1.45 GHz) per node 32 GiB RAM per node 125 PFlops Peak Power consumption: MW TOP 500: #1 Blue Gene/Q architecture 458,752 PowerPC A2 cores 16 cores (1.6 GHz) per node 16 GiB RAM per node 5D torus interconnect 5.8 PFlops Peak TOP 500: #13 Intel Xeon architecture 147,456 cores 16 cores (2.7 GHz) per node 32 GiB RAM per node Pruned tree interconnect 3.2 PFlops Peak TOP 500: #27 What is the problem? LBM Methods Ulrich Rüde

7 Only optimal algorithms qualify TERRA Thought experiment with O(N 2 ) algorithm Energy computer generation gigascale: 10 9 FLOPS terascale FLOPS petascale FLOPS exascale FLOPS desired problem size DoF=N energy estimate (kwh) 1 NJoule N 2 all-to-all communication Wh 10 min of LED light 278 kwh 2 weeks blow drying hair 278 GWh 1 month electricity for Hamburg 278 PWh 100 years world electricity production TerraNeo prototype (kwh) 0.13 Wh 0.03 kwh 27 kwh? At extreme scale: optimal complexity is a must! 7

8 Building block II: The Lagrangian View: Granular media simulations with the physics engine spherical particles 256 processors time steps runtime: 48 h (including data output) texture mapping, ray tracing Pöschel, T., & Schwager, T. (2005). Computational granular dynamics: models and algorithms. Springer Science & Business Media. 8

9 Lagrangian Particle Presentation Single particle described by state variables (position x, orientation φ, translational and angular velocity v and ω) a parameterization of its shape S (e.g. geometric primitive, composite object, or mesh) and its inertia properties (mass m, principle moments of inertia I xx, I yy and I zz ). The Newton-Euler equations of motion for rigid bodies describe the rate of change of the state variables: ẋ(t) = '(t) v(t) M('(t)) =!(t) v(t) Q('(t))!(t) f(s(t),t) (s(t),t)!(t) I('(t))!(t) 9

10 Hard Contacts alternative to the discrete element method Hard contacts require impulses, exhibit non-differentiable but continuous trajectories, contact reactions are defined implicitly in general, have non-unique solutions, and can be solved numerically by methods from two classes. ) measure differential inclusions 100% 80% 60% 40% 20% 0% -20% -1 soft contacts hard contacts Fig.: Bouncing ball with a soft and a hard contact model. Moreau, J., Panagiotopoulos P. (1988): Nonsmooth mechanics and applications, vol 302. Springer, Wien-New York Popa, C., Preclik, T., & UR (2014). Regularized solution of LCP problems with application to rigid body dynamics. Numerical Algorithms, Preclik, T. & UR (2015). Ultrascale simulations of non-smooth granular dynamics; Computational Particle Mechanics, DOI: /s Extreme Scale LBM Methods - Ulrich Rüde 10

11 Newton-Euler Equations for Rigid Bodies ẋ(t) v(t) = '(t) Q('(t))!(t) v(t) f(s(t),t) M('(t)) =!(t) (s(t),t)!(t) I('(t))!(t) Contact detection minimizing signed distance function Time-continuous non-penetration constraint for hard contacts Coulomb friction Parallelization via domain partitioning Parallel data structures for contact detection Protocol for synchronization (t) 0? n (t) 0 11

12 Nonlinear Complementarity and Time Stepping Non-penetration conditions Coulomb friction conditions continuous forces =0 + =0 0? n 0 k to k 2 apple µ n + 0? n 0 k v + tok 2 to = µ n v + to + 0? n 0 k v + tok 2 to = µ n v+ to k v + to k 2 =0 impulses =0 0? n 0 k to k 2 apple µ n + 0? n 0 k v + tok 2 to = µ n v + to discrete t + v0 n( ) 0? n 0 k to k 2 apple µ n k v 0 to( )k 2 to = µ n v 0 to( ) Signorini condition impact law friction cone condition frictional reaction opposes slip Moreau, J., Panagiotopoulos P. (1988): Nonsmooth mechanics and applications, vol 302. Springer, Wien-New York Popa, C., Preclik, T., & UR (2014). Regularized solution of LCP problems with application to rigid body dynamics. Numerical Algorithms, Preclik, T. & UR (2015). Ultrascale simulations of non-smooth granular dynamics; Computational Particle Mechanics, DOI: /s

13 Regularizing Multi-Contact Problems 13

14 Parallel Computation Key features of the parallelization: domain partitioning distribution of data contact detection synchronization protocol subdomain NBGS accumulators and corrections aggressive message aggregation nearest-neighbor communication Iglberger, K., & UR (2010). Massively parallel granular flow simulations with non-spherical particles. Computer Science-Research and Development, 25(1-2), Iglberger, K., & UR (2011). Large-scale rigid body simulations. Multibody System Dynamics, 25(1),

15 Shaker scenario with sharp edged hard objects sharp-edged particles with a diameter between 0.25 mm and 2 mm. 15

16 PE marble run - rigid objects in complex geometry Animation by Sebastian Eibl and Christian Godenschwager 16

17 Scaling Results Solver algorithmically not optimal for dense systems, hence cannot scale unconditionally, but is highly efficient in many cases of practical importance Strong and weak scaling results for a constant number of iterations performed on SuperMUC and Juqueen Largest ensembles computed non-spherical particles contacts granular gas: scaling results Breakup up of compute times on Erlangen RRZE Cluster Emmy av. time per time step and 1000 particles in s av. time per time step ( rst series) av. time per time step (second series) parallel e ciency (second series) number of nodes (b) Weak-scaling graph on the Juqueen supercomputer parallel e ciency 9.5% 25.9% 8.0% 12.6% (a) Time-step profile of the granular gas executed with = 20 processes on a single node. 25.8% 18.1% 8.3% 5.9% 22.7% 30.6% 16.5% 16.0% (b) Time-step profile of the granular gas executed with = 320 processes on 16 nodes. 17

18 Building Block III: Scalable Flow Simulations with the Lattice Boltzmann Method Lallemand, P., & Luo, L. S. (2000). Theory of the lattice Boltzmann method: Dispersion, dissipation, isotropy, Galilean invariance, and stability. Physical Review E, 61(6), Feichtinger, C., Donath, S., Köstler, H., Götz, J., & Rüde, U. (2011). WaLBerla: HPC software design for computational engineering simulations. Journal of Computational Science, 2(2), Extreme Scale LBM Methods - Ulrich Rüde 18

19 The Eulerian View : Lattice-Boltzmann Discretization in squares or cubes (cells) Particles Distribution Functions (PDF) in 2D: 9 numbers (2DQ9) in 3D: D3Q19 (alternatives D3Q27, etc) Repeat (many times) stream collide Extreme Scale LBM Methods - Ulrich Rüde 19

20 Basic Lattice Boltzmann Method Single Relaxation Time (SRT) Macroscopic quantities Equilibrium distribution function Extreme Scale LBM Methods - Ulrich Rüde

21 The stream step Move PDFs into neighboring cells Non-local part, Linear propagation to neighbors (stream step) Local part, Non-linear operator, (collide step) Extreme Scale LBM Methods - Ulrich Rüde 21

22 The collide step Compute new PDFs modeling molecular collisions Most collision operators can be expressed as Equilibrium function: non-linear, depending on the conserved moments,, and. Extreme Scale LBM Methods - Ulrich Rüde 22

23 Performance on Coronary Arteries Geometry Color coded proc assignment Godenschwager, C., Schornbaum, F., Bauer, M., Köstler, H., & UR (2013). A framework for hybrid parallel flow simulations with a trillion cells in complex geometries. In Proceedings of SC13: International Conference for High Performance Computing, Networking, Storage and Analysis (p. 35). ACM. Weak scaling 458,752 cores of JUQUEEN over a trillion (10 12 ) fluid lattice cells cell sizes 1.27µm diameter of red blood cells: 7µm cell updates per second 0.41 PFlops Strong scaling 32,768 cores of SuperMUC cell sizes of 0.1 mm 2.1 million fluid cells time steps per second Extreme Scale LBM Methods - Ulrich Rüde

24 Single Node Performance JUQUEEN SuperMUC vectorized optimized standard Pohl, T., Deserno, F., Thürey, N., UR, Lammers, P., Wellein, G., & Zeiser, T. (2004). Performance evaluation of parallel largescale lattice Boltzmann applications on three supercomputing architectures. Proceedings of the 2004 ACM/IEEE conference on Supercomputing (p. 21). IEEE Computer Society. Donath, S., Iglberger, K., Wellein, G., Zeiser, T., Nitsure, A., & UR (2008). Performance comparison of different parallel lattice Boltzmann implementations on multi-core multi-socket systems. International Journal of Computational Science and Engineering, 4(1), Extreme Scale LBM - Ulrich Rüde

25 Weak scaling for TRT lid driven cavity - uniform grids JUQUEEN 16 processes per node 4 threads per process cell updates per second (TLups) SuperMUC 4 processes per node 4 threads per process cell updates per second (TLups) Körner, C., Pohl, T., UR., Thürey, N., & Zeiser, T. (2006). Parallel lattice Boltzmann methods for CFD applications. In Numerical Solution of Partial Differential Equations on Parallel Computers (pp ). Springer Berlin Heidelberg. Feichtinger, C., Habich, J., Köstler, H., UR, & Aoki, T. (2015). Performance modeling and analysis of heterogeneous lattice Boltzmann simulations on CPU GPU clusters. Parallel Computing, 46, Extreme Scale LBM - Ulrich Rüde

26 Partitioning and Parallelization static block-level refinement ( forest of octrees) static load balancing DISK compact (KiB/MiB) binary MPI IO DISK allocation of block data ( grids) separation of domain partitioning from simulation (optional) 26

27 Flow through structure of thin crystals (filter) work with Jose Pedro Galache and Antonio Gil CMT-Motores Termicos, Universitat Politecnica de Valencia 27

28 Parallel AMR load balancing different views on domain partitioning 2:1 balanced grid (used for the LBM) distributed graph: nodes = blocks edges explicitly stored as < block ID, process rank > pairs forest of octrees: octrees are not explicitly stored, but implicitly defined via block IDs 28

29 AMR and Load Balancing with walberla Isaac, T., Burstedde, C., Wilcox, L. C., & Ghattas, O. (2015). Recursive algorithms for distributed forests of octrees. SIAM Journal on Scientific Computing, 37(5), C497-C531. Meyerhenke, H., Monien, B., & Sauerwald, T. (2009). A new diffusion-based multilevel algorithm for computing graph partitions. Journal of Parallel and Distributed Computing, 69(9), Schornbaum, F., & Rüde, U. (2016). Massively Parallel Algorithms for the Lattice Boltzmann Method on NonUniform Grids. SIAM Journal on Scientific Computing, 38(2), C96-C126. Extreme Scale LBM Methods - Ulrich Rüde 29

30 AMR Performance Benchmark Environments: JUQUEEN (5.0 PFLOP/s) Blue Gene/Q, 459K cores, 1 GB/core compiler: IBM XL / IBM MPI SuperMUC (2.9 PFLOP/s) Intel Xeon, 147K cores, 2 GB/core compiler: Intel XE / IBM MPI Benchmark (LBM D3Q19 TRT): avg. blocks/process (max. blocks/proc.) level initially after refresh after load balance (1) (1) (1) (1) (9) (1) (2) (11) (4) (4) (16) (4) Extreme Scale LBM Methods - Ulrich Rüde 30

31 AMR Performance Benchmark Environments: JUQUEEN (5.0 PFLOP/s) Blue Gene/Q, 459K cores, 1 GB/core compiler: IBM XL / IBM MPI SuperMUC (2.9 PFLOP/s) Intel Xeon, 147K cores, 2 GB/core compiler: Intel XE / IBM MPI Benchmark (LBM D3Q19 TRT): during this refresh process all cells on the finest level are coarsened and coarsen refine the same amount of fine cells is created by splitting coarser cells 72 % of all cells change their size Extreme Scale LBM Methods - Ulrich Rüde 31

32 AMR Performance JUQUEEN space filling curve: Morton billion cells seconds billion cells 14 billion cells hybrid MPI+OpenMP version with SMP 1 process 2 cores 8 threads #cells per core 31, , , , ,752 cores Extreme Scale LBM Methods - Ulrich Rüde 32

33 AMR Performance JUQUEEN diffusion load balancing billion cells seconds billion cells 14 billion cells time almost independent of #processes! #cells per core 31, , , , ,752 cores Extreme Scale LBM Methods - Ulrich Rüde 33

34 Pore scale resolved flow in porous media Direct numerical simulation of flow through sphere packings Beetstra, R., Van der Hoef, M. A., & Kuipers, J. A. M. (2007). Drag force of intermediate Reynolds number flow past monoand bidisperse arrays of spheres. AIChE Journal, 53(2), Tenneti, S., Garg, R., & Subramaniam, S. (2011). Drag law for monodisperse gas solid systems using particle-resolved direct numerical simulation of flow past fixed assemblies of spheres. International journal of multiphase flow, 37(9),

35 Flow field and vorticity 2D slice visualized Domain size: Re = 300 Volume fraction:

36 Macroscopic drag correlation Finally, the drag correlation reads Average absolute percentage error: 9.7 % Bogner, S., Mohanty, S., & UR (2015). Drag correlation for dilute and moderately dense fluid-particle systems using the lattice Boltzmann method, International Journal of Multiphase Flow 68,

37 Setup of random spherical structure for porous media with the PE Massively Parallel Multiphysics Simulations - Ulrich Rüde

38 Flow over porous structure Porosity Setup Height of the channel Porosity Velocity Profile Streamwise velocity Pore geometry and streamlines Fattahi, E., Waluga, C., Wohlmuth, B., & Rüde, U. (2015). Large scale lattice Boltzmann simulation for the coupling of free and porous media flow. arxiv preprint arxiv: To appear in LNCS, Proceedings of HPCSE Massively Parallel Multiphysics Simulations - Ulrich Rüde 38

39 Turbulent flow over a permeable region Re 3000 direct numerical simulation Volume rendering of velocity magnitude Periodic in X and Y direction I10 cluster, 7x32x19=4256 core hours 1,300,000 timesteps 8 times more time steps on the finest level 39

40 Multi-Physics Simulations for Particulate Flows Parallel Coupling with walberla and PE Ladd, A. J. (1994). Numerical simulations of particulate suspensions via a discretized Boltzmann equation. Part 1. Theoretical foundation. Journal of Fluid Mechanics, 271(1), Tenneti, S., & Subramaniam, S. (2014). Particle-resolved direct numerical simulation for gas-solid flow model development. Annual Review of Fluid Mechanics, 46, Bartuschat, D., Fischermeier, E., Gustavsson, K., & UR (2016). Two computational models for simulating the tumbling motion of elongated particles in fluids. Computers & Fluids, 127,

41 Fluid-Structure Interaction direct simulation of Particle Laden Flows (4-way coupling) Götz, J., Iglberger, K., Stürmer, M., & UR (2010). Direct numerical simulation of particulate flows on processor cores. In Proceedings of Supercomputing 2010, IEEE Computer Society. Götz, J., Iglberger, K., Feichtinger, C., Donath, S., & UR (2010). Coupling multibody dynamics and computational fluid dynamics on 8192 processor cores. Parallel Computing, 36(2),

42 Simulation of sediment transport 42

43 Heterogenous CPU-GPU Simulation Particles: 31250, Domain: 400x400x200, Timesteps: Devices: 2 x M Intel Westmere, Runtime: 17.5 h Fluidized Beds: Direct numerical simulation fully resolved particles Fluid-structureinteraction 4-way-coupling C. Feichtinger, J. Habich, H. Köstler, U. Rüde, T. Aoki, Performance modeling and analysis of heterogeneous lattice Boltzmann simulations on CPU GPU clusters, Parallel Computing, Volume 46, July 2015, Pages 1-13 C. Feichtinger, J. Habich, H. Köstler, G. Hager, U. Rüde, G. Wellein, A flexible Patch-based lattice Boltzmann parallelization approach for heterogeneous GPU CPU clusters, Parallel Computing, Volume 37, Issue 9, September 2011, Pages

44 Weak Scaling experiment Götz, J., Iglberger, K., Stürmer, M., & Rüde, U. (2010, November). Direct numerical simulation of particulate flows on processor cores. In Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis (pp. 1-11). IEEE Computer Society. Efficiency Jugene Blue Gene/P Jülich Supercomputer Center 40x40x40 lattice cells per core 80x80x80 lattice cells per core Number of Cores Largest simulation on Jugene: 8 Trillion (10 12 ) variables per time step (LBM alone) Scaling 64 to cores Densely packed particles lattice cells rigid spherical objects Massively Parallel Multiphysics Simulations - Ulrich Rüde 44

45 Building Block IV (electrostatics) Direct numerical simulation of charged particles in flow Positive and negatively charged particles in flow subjected to transversal electric field Masilamani, K., Ganguly, S., Feichtinger, C., & UR (2011). Hybrid lattice-boltzmann and finite-difference simulation of electroosmotic flow in a microchannel. Fluid Dynamics Research, 43(2), Bartuschat, D., Ritter, D., & UR (2012). Parallel multigrid for electrokinetic simulation in particle-fluid flows. In High Performance Computing and Simulation (HPCS), 2012 International Conference on (pp ). IEEE. Bartuschat, D. & UR (2015). Parallel Multiphysics Simulations of Charged Particles in Microfluidic Flows, Journal of Computational Science, Volume 8, May 2015, Pages

46 6-way coupling charge distribution velocity BCs Finite volumes MG treat BCs V-cycle iter at. object motion hydrodynam. force LBM treat BCs stream-collide step electrostat. force Newtonian mechanics collision response object distance correction force Lubrication correction 46

47 Separation experiment 240 time steps fully 6-way coupled simulation 400 sec on SuperMuc weak scaling up to cores 7.1 Mio particles Total runtimes [] Number of nodes LBM Map Lubr HydrF pe MG SetRHS PtCm ElectF 10 3 MFLUPS (LBM) Number of nodes 47 LBM Perform. MG Perform MLUPS (MG)

48 Building Block V Volume of Fluids Method for Free Surface Flows joint work with Regina Ammer, Simon Bogner, Martin Bauer, Daniela Anderl, Nils Thürey, Stefan Donath, Thomas Pohl, C Körner, A. Delgado Körner, C., Thies, M., Hofmann, T., Thürey, N., & UR. (2005). Lattice Boltzmann model for free surface flow for modeling foaming. Journal of Statistical Physics, 121(1-2), Donath, S., Feichtinger, C., Pohl, T., Götz, J., & UR. (2010). A Parallel Free Surface Lattice Boltzmann Method for Large- Scale Applications. Parallel Computational Fluid Dynamics: Recent Advances and Future Directions, 318. Anderl, D., Bauer, M., Rauh, C., UR, & Delgado, A. (2014). Numerical simulation of adsorption and bubble interaction in protein foams using a lattice Boltzmann method. Food & function, 5(4), Bogner, S., Ammer, R., & Rüde, U. (2015). Boundary conditions for free interfaces with the lattice Boltzmann method. Journal of Computational Physics, 297,

49 Free Surface Flows Volume-of-Fluids like approach Flag field: Compute only in fluid Special free surface conditions in interface cells Reconstruction of curvature for surface tension 49

50 Simulation of Metal Foams Example application: Engineering: metal foam simulations Based on LBM: Free surfaces Surface tension Disjoining pressure to stabilize thin liquid films Parallelization with MPI and load balancing Other applications: Food processing Fuel cells Körner, C., Thies, M., Hofmann, T., Thürey, N., & UR (2005). Lattice Boltzmann model for free surface flow for modeling foaming. Journal of Statistical Physics, 121(1-2), Donath, S., Mecke, K., Rabha, S., Buwa, V., & UR (2011). Verification of surface tension in the parallel free surface lattice Boltzmann method in walberla. Computers & Fluids, 45(1), Thürey, N., &UR. (2009). Stable free surface flows with the lattice Boltzmann method on adaptively coarsened grids. Computing and Visualization in Science, 12(5),

51 Simulation for hygiene products (for Procter&Gamble) capillary pressure surface tension inclination contact angle 51

52 Additive Manufacturing Fast Electron Beam Melting Ammer, R., Markl, M., Ljungblad, U., Körner, C., & UR (2014). Simulating fast electron beam melting with a parallel thermal free surface lattice Boltzmann method. Computers & Mathematics with Applications, 67(2), Ammer, R., UR, Markl, M., Jüchter V., & Körner, C. (2014). Validation experiments for LBM simulations of electron beam melting. International Journal of Modern Physics C. 52

53 Simulation of Electron Beam Melting Simulating powder bed generation using the PE framework High speed camera shows melting step for manufacturing a hollow cylinder WaLBerla Simulation 53

54 Conclusions 54

55 Research in Computational Science is done by teams Harald Köstler Christian Godenschwager Kristina Pickl Regina Ammer Simon Bogner Ehsan Fattahi Florian Schornbaum Sebastian Kuckuk Christoph Rettinger Dominik Bartuschat Martin Bauer Christian Kuschel 55

56 Thank you for your attention! Videos, preprints, slides at 56

Lattice Boltzmann Methods on the way to exascale

Lattice Boltzmann Methods on the way to exascale Lattice Boltzmann Methods on the way to exascale Ulrich Rüde (LSS Erlangen, ulrich.ruede@fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de HIGH PERFORMANCE

More information

The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms

The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms Harald Köstler, Uli Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

Peta-Scale Simulations with the HPC Software Framework walberla:

Peta-Scale Simulations with the HPC Software Framework walberla: Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,

More information

simulation framework for piecewise regular grids

simulation framework for piecewise regular grids WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler

More information

Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm

Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm March 17 March 21, 2014 Florian Schornbaum, Martin Bauer, Simon Bogner Chair for System Simulation Friedrich-Alexander-Universität

More information

walberla: Developing a Massively Parallel HPC Framework

walberla: Developing a Massively Parallel HPC Framework walberla: Developing a Massively Parallel HPC Framework SIAM CS&E 2013, Boston February 26, 2013 Florian Schornbaum*, Christian Godenschwager*, Martin Bauer*, Matthias Markl, Ulrich Rüde* *Chair for System

More information

(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more

(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more Parallel Free-Surface Extension of the Lattice-Boltzmann Method A Lattice-Boltzmann Approach for Simulation of Two-Phase Flows Stefan Donath (LSS Erlangen, stefan.donath@informatik.uni-erlangen.de) Simon

More information

Massively Parallel Phase Field Simulations using HPC Framework walberla

Massively Parallel Phase Field Simulations using HPC Framework walberla Massively Parallel Phase Field Simulations using HPC Framework walberla SIAM CSE 2015, March 15 th 2015 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer, Harald Köstler and Ulrich

More information

Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method

Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method June 21, 2011 Introduction Free Surface LBM Liquid-Gas-Solid Flows Parallel Computing Examples and More References Fig. Simulation

More information

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla

Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla SIAM PP 2016, April 13 th 2016 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer,

More information

Software and Performance Engineering for numerical codes on GPU clusters

Software and Performance Engineering for numerical codes on GPU clusters Software and Performance Engineering for numerical codes on GPU clusters H. Köstler International Workshop of GPU Solutions to Multiscale Problems in Science and Engineering Harbin, China 28.7.2010 2 3

More information

Large scale Imaging on Current Many- Core Platforms

Large scale Imaging on Current Many- Core Platforms Large scale Imaging on Current Many- Core Platforms SIAM Conf. on Imaging Science 2012 May 20, 2012 Dr. Harald Köstler Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen,

More information

Reconstruction of Trees from Laser Scan Data and further Simulation Topics

Reconstruction of Trees from Laser Scan Data and further Simulation Topics Reconstruction of Trees from Laser Scan Data and further Simulation Topics Helmholtz-Research Center, Munich Daniel Ritter http://www10.informatik.uni-erlangen.de Overview 1. Introduction of the Chair

More information

Performance and Software-Engineering Considerations for Massively Parallel Simulations

Performance and Software-Engineering Considerations for Massively Parallel Simulations Performance and Software-Engineering Considerations for Massively Parallel Simulations Ulrich Rüde (ruede@cs.fau.de) Ben Bergen, Frank Hülsemann, Christoph Freundl Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

A Contact Angle Model for the Parallel Free Surface Lattice Boltzmann Method in walberla Stefan Donath (stefan.donath@informatik.uni-erlangen.de) Computer Science 10 (System Simulation) University of Erlangen-Nuremberg

More information

A Python extension for the massively parallel framework walberla

A Python extension for the massively parallel framework walberla A Python extension for the massively parallel framework walberla PyHPC at SC 14, November 17 th 2014 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler

More information

Towards PetaScale Computational Science

Towards PetaScale Computational Science Towards PetaScale Computational Science U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with many Lehrstuhl für Informatik 10 (Systemsimulation) Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de

More information

Numerical Algorithms on Multi-GPU Architectures

Numerical Algorithms on Multi-GPU Architectures Numerical Algorithms on Multi-GPU Architectures Dr.-Ing. Harald Köstler 2 nd International Workshops on Advances in Computational Mechanics Yokohama, Japan 30.3.2010 2 3 Contents Motivation: Applications

More information

Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method

Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method G. Wellein, T. Zeiser, G. Hager HPC Services Regional Computing Center A. Nitsure, K. Iglberger, U. Rüde Chair for System

More information

Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures

Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Jan Treibig, Simon Hausmann, Ulrich Ruede Zusammenfassung The Lattice Boltzmann method (LBM) is a well established algorithm

More information

Towards Exa-Scale: Computing with Millions of Cores

Towards Exa-Scale: Computing with Millions of Cores Towards Exa-Scale: Computing with Millions of Cores U. Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Informatik 10 (Systemsimulation) Excellence Cluster Engineering of Advanced Materials Universität

More information

τ-extrapolation on 3D semi-structured finite element meshes

τ-extrapolation on 3D semi-structured finite element meshes τ-extrapolation on 3D semi-structured finite element meshes European Multi-Grid Conference EMG 2010 Björn Gmeiner Joint work with: Tobias Gradl, Ulrich Rüde September, 2010 Contents The HHG Framework τ-extrapolation

More information

Automatic Generation of Algorithms and Data Structures for Geometric Multigrid. Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014

Automatic Generation of Algorithms and Data Structures for Geometric Multigrid. Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014 Automatic Generation of Algorithms and Data Structures for Geometric Multigrid Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014 Introduction Multigrid Goal: Solve a partial differential

More information

Simulation of moving Particles in 3D with the Lattice Boltzmann Method

Simulation of moving Particles in 3D with the Lattice Boltzmann Method Simulation of moving Particles in 3D with the Lattice Boltzmann Method, Nils Thürey, Christian Feichtinger, Hans-Joachim Schmid Chair for System Simulation University Erlangen/Nuremberg Chair for Particle

More information

Direct Numerical Simulation of Particulate Flows on Processor Cores

Direct Numerical Simulation of Particulate Flows on Processor Cores 1 Direct Numerical Simulation of Particulate Flows on 294 912 Processor Cores Jan Götz *, Klaus Iglberger *, Markus Stürmer *, and Ulrich Rüde *,** * Chair for System Simulation, University Erlangen-Nuremberg,

More information

Massively Parallel Finite Element Simulations with deal.ii

Massively Parallel Finite Element Simulations with deal.ii Massively Parallel Finite Element Simulations with deal.ii Timo Heister, Texas A&M University 2012-02-16 SIAM PP2012 joint work with: Wolfgang Bangerth, Carsten Burstedde, Thomas Geenen, Martin Kronbichler

More information

HIGH PERFORMANCE COMPUTATION (HPC) FOR THE

HIGH PERFORMANCE COMPUTATION (HPC) FOR THE HIGH PERFORMANCE COMPUTATION (HPC) FOR THE DEVELOPMENT OF FLUIDIZED BED TECHNOLOGIES FOR BIOMASS GASIFICATION AND CO2 CAPTURE P. Fede, H. Neau, O. Simonin Université de Toulouse; INPT, UPS ; IMFT ; 31400

More information

Sustainability and Efficiency for Simulation Software in the Exascale Era

Sustainability and Efficiency for Simulation Software in the Exascale Era Sustainability and Efficiency for Simulation Software in the Exascale Era Dominik Thönnes, Ulrich Rüde, Nils Kohl Chair for System Simulation, University of Erlangen-Nürnberg March 09, 2018 SIAM Conference

More information

Highly efficient «on-the-fly» data processing using the open-source library CPPPO

Highly efficient «on-the-fly» data processing using the open-source library CPPPO Highly efficient «on-the-fly» data processing using the open-source library CPPPO Graz University of Technology, DCS Computing GmbH Federico Municchi, Stefan Radl, Christoph Goniva April 7 2016, Workshop

More information

Lehrstuhl für Informatik 10 (Systemsimulation)

Lehrstuhl für Informatik 10 (Systemsimulation) FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) On the Resource Requirements of

More information

CFD in COMSOL Multiphysics

CFD in COMSOL Multiphysics CFD in COMSOL Multiphysics Christian Wollblad Copyright 2017 COMSOL. Any of the images, text, and equations here may be copied and modified for your own internal use. All trademarks are the property of

More information

Adaptive Hierarchical Grids with a Trillion Tetrahedra

Adaptive Hierarchical Grids with a Trillion Tetrahedra Adaptive Hierarchical Grids with a Trillion Tetrahedra Tobias Gradl, Björn Gmeiner and U. Rüde (LSS Erlangen, ruede@cs.fau.de) in collaboration with many more Lehrstuhl für Informatik 10 (Systemsimulation)

More information

Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids

Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Sebastian Kuckuk FAU Erlangen-Nürnberg 18.01.2016 HiStencils 2016, Prague, Czech Republic Outline Outline Scope

More information

High Performance Computing

High Performance Computing High Performance Computing ADVANCED SCIENTIFIC COMPUTING Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich

More information

Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs)

Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) OBJECTIVE FLUID SIMULATIONS Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) The basic objective of the project is the implementation of the paper Stable Fluids (Jos Stam, SIGGRAPH 99). The final

More information

Generic finite element capabilities for forest-of-octrees AMR

Generic finite element capabilities for forest-of-octrees AMR Generic finite element capabilities for forest-of-octrees AMR Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

CFD MODELING FOR PNEUMATIC CONVEYING

CFD MODELING FOR PNEUMATIC CONVEYING CFD MODELING FOR PNEUMATIC CONVEYING Arvind Kumar 1, D.R. Kaushal 2, Navneet Kumar 3 1 Associate Professor YMCAUST, Faridabad 2 Associate Professor, IIT, Delhi 3 Research Scholar IIT, Delhi e-mail: arvindeem@yahoo.co.in

More information

Lattice Boltzmann with CUDA

Lattice Boltzmann with CUDA Lattice Boltzmann with CUDA Lan Shi, Li Yi & Liyuan Zhang Hauptseminar: Multicore Architectures and Programming Page 1 Outline Overview of LBM An usage of LBM Algorithm Implementation in CUDA and Optimization

More information

Support for Multi physics in Chrono

Support for Multi physics in Chrono Support for Multi physics in Chrono The Story Ahead Overview of multi physics strategy in Chrono Summary of handling rigid/flexible body dynamics using Lagrangian approach Summary of handling fluid, and

More information

Solving Partial Differential Equations on Overlapping Grids

Solving Partial Differential Equations on Overlapping Grids **FULL TITLE** ASP Conference Series, Vol. **VOLUME**, **YEAR OF PUBLICATION** **NAMES OF EDITORS** Solving Partial Differential Equations on Overlapping Grids William D. Henshaw Centre for Applied Scientific

More information

LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS

LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS 14 th European Conference on Mixing Warszawa, 10-13 September 2012 LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS Felix Muggli a, Laurent Chatagny a, Jonas Lätt b a Sulzer Markets & Technology

More information

Hierarchical Hybrid Grids

Hierarchical Hybrid Grids Hierarchical Hybrid Grids IDK Summer School 2012 Björn Gmeiner, Ulrich Rüde July, 2012 Contents Mantle convection Hierarchical Hybrid Grids Smoothers Geometric approximation Performance modeling 2 Mantle

More information

Forces on particles and bubbles Literature

Forces on particles and bubbles Literature Lecture 1: Forces on particles and bubbles In order to describe the motion of particles and bubbles in fluids it is necessary to consider all relevant forces, external and fluid dynamic forces. Starting

More information

arxiv: v1 [cs.pf] 5 Dec 2011

arxiv: v1 [cs.pf] 5 Dec 2011 Performance engineering for the Lattice Boltzmann method on GPGPUs: Architectural requirements and performance results J. Habich a, C. Feichtinger b, H. Köstler b, G. Hager a, G. Wellein a,b a Erlangen

More information

Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics

Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics I. Pantle Fachgebiet Strömungsmaschinen Karlsruher Institut für Technologie KIT Motivation

More information

Algorithms, System and Data Centre Optimisation for Energy Efficient HPC

Algorithms, System and Data Centre Optimisation for Energy Efficient HPC 2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy

More information

Development of Hybrid Fluid Jet / Float Polishing Process

Development of Hybrid Fluid Jet / Float Polishing Process COMSOL Conference - Tokyo 2013 Development of Hybrid Fluid Jet / Float Polishing Process A. Beaucamp, Y. Namba Dept. of Mechanical Engineering, Chubu University, Japan Zeeko LTD, United Kingdom Research

More information

Final drive lubrication modeling

Final drive lubrication modeling Final drive lubrication modeling E. Avdeev a,b 1, V. Ovchinnikov b a Samara University, b Laduga Automotive Engineering Abstract. In this paper we describe the method, which is the composition of finite

More information

Applications of ICFD /SPH Solvers by LS-DYNA to Solve Water Splashing Impact to Automobile Body. Abstract

Applications of ICFD /SPH Solvers by LS-DYNA to Solve Water Splashing Impact to Automobile Body. Abstract Applications of ICFD /SPH Solvers by LS-DYNA to Solve Water Splashing Impact to Automobile Body George Wang (1 ), Kevin Gardner (3), Eric DeHoff (1), Facundo del Pin (2), Inaki Caldichoury (2), Edouard

More information

Free Surface Flows with Moving and Deforming Objects for LBM

Free Surface Flows with Moving and Deforming Objects for LBM Free Surface Flows with Moving and Deforming Objects for LBM Nils Thürey, Klaus Iglberger, Ulrich Rüde Institute for System Simulation (LSS), University of Erlangen-Nuremberg, Cauerstr. 6, 91058 Erlangen,

More information

Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures

Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Dirk Ribbrock, Markus Geveler, Dominik Göddeke, Stefan Turek Angewandte Mathematik, Technische Universität Dortmund

More information

Free Surface Lattice-Boltzmann fluid simulations. with and without level sets.

Free Surface Lattice-Boltzmann fluid simulations. with and without level sets. ree Surface Lattice-Boltzmann fluid simulations with and without level sets Nils Thürey, Ulrich Rüde University of Erlangen-Nuremberg System Simulation roup Cauerstr. 6, 91054 Erlangen, ermany Email: Nils.Thuerey@cs.fau.de

More information

Preliminary Spray Cooling Simulations Using a Full-Cone Water Spray

Preliminary Spray Cooling Simulations Using a Full-Cone Water Spray 39th Dayton-Cincinnati Aerospace Sciences Symposium Preliminary Spray Cooling Simulations Using a Full-Cone Water Spray Murat Dinc Prof. Donald D. Gray (advisor), Prof. John M. Kuhlman, Nicholas L. Hillen,

More information

Towards Generating Solvers for the Simulation of non-newtonian Fluids. Harald Köstler, Sebastian Kuckuk FAU Erlangen-Nürnberg

Towards Generating Solvers for the Simulation of non-newtonian Fluids. Harald Köstler, Sebastian Kuckuk FAU Erlangen-Nürnberg Towards Generating Solvers for the Simulation of non-newtonian Fluids Harald Köstler, Sebastian Kuckuk FAU Erlangen-Nürnberg 22.12.2015 Outline Outline Scope and Motivation Project ExaStencils The Application

More information

Performance Metrics of a Parallel Three Dimensional Two-Phase DSMC Method for Particle-Laden Flows

Performance Metrics of a Parallel Three Dimensional Two-Phase DSMC Method for Particle-Laden Flows Performance Metrics of a Parallel Three Dimensional Two-Phase DSMC Method for Particle-Laden Flows Benzi John* and M. Damodaran** Division of Thermal and Fluids Engineering, School of Mechanical and Aerospace

More information

Particleworks: Particle-based CAE Software fully ported to GPU

Particleworks: Particle-based CAE Software fully ported to GPU Particleworks: Particle-based CAE Software fully ported to GPU Introduction PrometechVideo_v3.2.3.wmv 3.5 min. Particleworks Why the particle method? Existing methods FEM, FVM, FLIP, Fluid calculation

More information

Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics

Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Masanori Hashiguchi 1 1 Keisoku Engineering System Co., Ltd. 1-9-5 Uchikanda, Chiyoda-ku,

More information

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy

Enzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2

More information

ORAP Forum October 10, 2013

ORAP Forum October 10, 2013 Towards Petaflop simulations of core collapse supernovae ORAP Forum October 10, 2013 Andreas Marek 1 together with Markus Rampp 1, Florian Hanke 2, and Thomas Janka 2 1 Rechenzentrum der Max-Planck-Gesellschaft

More information

Optimizing Bio-Inspired Flow Channel Design on Bipolar Plates of PEM Fuel Cells

Optimizing Bio-Inspired Flow Channel Design on Bipolar Plates of PEM Fuel Cells Excerpt from the Proceedings of the COMSOL Conference 2010 Boston Optimizing Bio-Inspired Flow Channel Design on Bipolar Plates of PEM Fuel Cells James A. Peitzmeier *1, Steven Kapturowski 2 and Xia Wang

More information

Direct Numerical Simulation of Turbulent Boundary Layers at High Reynolds Numbers.

Direct Numerical Simulation of Turbulent Boundary Layers at High Reynolds Numbers. Direct Numerical Simulation of Turbulent Boundary Layers at High Reynolds Numbers. G. Borrell, J.A. Sillero and J. Jiménez, Corresponding author: guillem@torroja.dmt.upm.es School of Aeronautics, Universidad

More information

CGT 581 G Fluids. Overview. Some terms. Some terms

CGT 581 G Fluids. Overview. Some terms. Some terms CGT 581 G Fluids Bedřich Beneš, Ph.D. Purdue University Department of Computer Graphics Technology Overview Some terms Incompressible Navier-Stokes Boundary conditions Lagrange vs. Euler Eulerian approaches

More information

SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS

SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22, 2012 SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC

More information

Lattice Boltzmann Liquid Simulation with Moving Objects on Graphics Hardware

Lattice Boltzmann Liquid Simulation with Moving Objects on Graphics Hardware Lattice Boltzmann Liquid Simulation with Moving Objects on Graphics Hardware Duncan Clough Supervised by: James Gain, Michelle Kuttel 1 Introduction Fluid simulation is a common element in recent feature

More information

LARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS

LARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS ECCOMAS Congress 2016 VII European Congress on Computational Methods in Applied Sciences and Engineering M. Papadrakakis, V. Papadopoulos, G. Stefanou, V. Plevris (eds.) Crete Island, Greece, 5 10 June

More information

A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla

A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler, and Ulrich

More information

High performance computing and numerical modeling

High performance computing and numerical modeling High performance computing and numerical modeling Volker Springel Plan for my lectures Lecture 1: Collisional and collisionless N-body dynamics Lecture 2: Gravitational force calculation Lecture 3: Basic

More information

ENERGY-224 Reservoir Simulation Project Report. Ala Alzayer

ENERGY-224 Reservoir Simulation Project Report. Ala Alzayer ENERGY-224 Reservoir Simulation Project Report Ala Alzayer Autumn Quarter December 3, 2014 Contents 1 Objective 2 2 Governing Equations 2 3 Methodolgy 3 3.1 BlockMesh.........................................

More information

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation) FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) walberla: Visualization of Fluid

More information

FOUR WHAT S NEW IN THIS VERSION? 4.1 FLOW-3D Usability CHAPTER

FOUR WHAT S NEW IN THIS VERSION? 4.1 FLOW-3D Usability CHAPTER CHAPTER FOUR WHAT S NEW IN THIS VERSION? FLOW-3D v11.2.0 continues to streamline engineers simulation workflows by enabling them to more quickly set up simulations, avoid common errors, identify and enter

More information

Introduction to C omputational F luid Dynamics. D. Murrin

Introduction to C omputational F luid Dynamics. D. Murrin Introduction to C omputational F luid Dynamics D. Murrin Computational fluid dynamics (CFD) is the science of predicting fluid flow, heat transfer, mass transfer, chemical reactions, and related phenomena

More information

Dynamics in Maya. Gary Monheit Alias Wavefront PHYSICALLY BASED MODELING SH1 SIGGRAPH 97 COURSE NOTES

Dynamics in Maya. Gary Monheit Alias Wavefront PHYSICALLY BASED MODELING SH1 SIGGRAPH 97 COURSE NOTES Dynamics in Maya Gary Monheit Alias Wavefront SH1 Dynamics in Maya Overall Requirements Architecture and Features Animations SH2 Overall Requirements Why Dynamics? Problems with traditional animation techniques

More information

A Peta-scale LES (Large-Eddy Simulation) for Turbulent Flows Based on Lattice Boltzmann Method

A Peta-scale LES (Large-Eddy Simulation) for Turbulent Flows Based on Lattice Boltzmann Method GTC (GPU Technology Conference) 2013, San Jose, 2013, March 20 A Peta-scale LES (Large-Eddy Simulation) for Turbulent Flows Based on Lattice Boltzmann Method Takayuki Aoki Global Scientific Information

More information

Driven Cavity Example

Driven Cavity Example BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square

More information

Phys 113 Final Project

Phys 113 Final Project Phys 113 Final Project LBM model for capillary fluid modelling Andrey Sushko Introduction Lattice Boltzmann Method (LMB) models have recently emerged as a highly effective means of modelling complex fluid

More information

Parallel High-Order Geometric Multigrid Methods on Adaptive Meshes for Highly Heterogeneous Nonlinear Stokes Flow Simulations of Earth s Mantle

Parallel High-Order Geometric Multigrid Methods on Adaptive Meshes for Highly Heterogeneous Nonlinear Stokes Flow Simulations of Earth s Mantle ICES Student Forum The University of Texas at Austin, USA November 4, 204 Parallel High-Order Geometric Multigrid Methods on Adaptive Meshes for Highly Heterogeneous Nonlinear Stokes Flow Simulations of

More information

Abstract. 1 Introduction

Abstract. 1 Introduction On the Use of Pore-Scale Computational Models for Two-Phase Porous-media Flows M.A. Celia, P.C Reeves, H.K. Dahle Environmental Engineering and Water Resources Program, Department of Civil Engineering

More information

LS-DYNA 980 : Recent Developments, Application Areas and Validation Process of the Incompressible fluid solver (ICFD) in LS-DYNA.

LS-DYNA 980 : Recent Developments, Application Areas and Validation Process of the Incompressible fluid solver (ICFD) in LS-DYNA. 12 th International LS-DYNA Users Conference FSI/ALE(1) LS-DYNA 980 : Recent Developments, Application Areas and Validation Process of the Incompressible fluid solver (ICFD) in LS-DYNA Part 1 Facundo Del

More information

High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters

High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters SIAM PP 2014 High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters C. Riesinger, A. Bakhtiari, M. Schreiber Technische Universität München February 20, 2014

More information

I/O at JSC. I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O. Wolfgang Frings

I/O at JSC. I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O. Wolfgang Frings Mitglied der Helmholtz-Gemeinschaft I/O at JSC I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O Wolfgang Frings W.Frings@fz-juelich.de Jülich Supercomputing

More information

NIA CFD Futures Conference Hampton, VA; August 2012

NIA CFD Futures Conference Hampton, VA; August 2012 Petascale Computing and Similarity Scaling in Turbulence P. K. Yeung Schools of AE, CSE, ME Georgia Tech pk.yeung@ae.gatech.edu NIA CFD Futures Conference Hampton, VA; August 2012 10 2 10 1 10 4 10 5 Supported

More information

Simulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg

Simulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg Simulation in Computer Graphics Deformable Objects Matthias Teschner Computer Science Department University of Freiburg Outline introduction forces performance collision handling visualization University

More information

SPH: Why and what for?

SPH: Why and what for? SPH: Why and what for? 4 th SPHERIC training day David Le Touzé, Fluid Mechanics Laboratory, Ecole Centrale de Nantes / CNRS SPH What for and why? How it works? Why not for everything? Duality of SPH SPH

More information

Realtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research

Realtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research 1 Realtime Water Simulation on GPU Nuttapong Chentanez NVIDIA Research 2 3 Overview Approaches to realtime water simulation Hybrid shallow water solver + particles Hybrid 3D tall cell water solver + particles

More information

Particle Tracing Module

Particle Tracing Module Particle Tracing Module Particle Tracing Module Released with version 4.2a in October 2011 Add-on to COMSOL Multiphysics Combines with any COMSOL Multiphysics Module Particle Tracing Particle tracing

More information

Efficient Imaging Algorithms on Many-Core Platforms

Efficient Imaging Algorithms on Many-Core Platforms Efficient Imaging Algorithms on Many-Core Platforms H. Köstler Dagstuhl, 22.11.2011 Contents Imaging Applications HDR Compression performance of PDE-based models Image Denoising performance of patch-based

More information

Part I: Theoretical Background and Integration-Based Methods

Part I: Theoretical Background and Integration-Based Methods Large Vector Field Visualization: Theory and Practice Part I: Theoretical Background and Integration-Based Methods Christoph Garth Overview Foundations Time-Varying Vector Fields Numerical Integration

More information

Coupling of STAR-CCM+ to Other Theoretical or Numerical Solutions. Milovan Perić

Coupling of STAR-CCM+ to Other Theoretical or Numerical Solutions. Milovan Perić Coupling of STAR-CCM+ to Other Theoretical or Numerical Solutions Milovan Perić Contents The need to couple STAR-CCM+ with other theoretical or numerical solutions Coupling approaches: surface and volume

More information

MESHLESS SOLUTION OF INCOMPRESSIBLE FLOW OVER BACKWARD-FACING STEP

MESHLESS SOLUTION OF INCOMPRESSIBLE FLOW OVER BACKWARD-FACING STEP Vol. 12, Issue 1/2016, 63-68 DOI: 10.1515/cee-2016-0009 MESHLESS SOLUTION OF INCOMPRESSIBLE FLOW OVER BACKWARD-FACING STEP Juraj MUŽÍK 1,* 1 Department of Geotechnics, Faculty of Civil Engineering, University

More information

Direct Numerical Simulation of a Low Pressure Turbine Cascade. Christoph Müller

Direct Numerical Simulation of a Low Pressure Turbine Cascade. Christoph Müller Low Pressure NOFUN 2015, Braunschweig, Overview PostProcessing Experimental test facility Grid generation Inflow turbulence Conclusion and slide 2 / 16 Project Scale resolving Simulations give insight

More information

Computational Fluid Dynamic Hydraulic Characterization: G3 Cube vs. Dolos Armour Unit. IS le Roux, WJS van der Merwe & CL de Wet

Computational Fluid Dynamic Hydraulic Characterization: G3 Cube vs. Dolos Armour Unit. IS le Roux, WJS van der Merwe & CL de Wet Computational Fluid Dynamic Hydraulic Characterization: G3 Cube vs. Dolos Armour Unit IS le Roux, WJS van der Merwe & CL de Wet Presentation Outline Scope. Assumptions and boundary values. Numerical mesh.

More information

Large Scale Parallel Lattice Boltzmann Model of Dendritic Growth

Large Scale Parallel Lattice Boltzmann Model of Dendritic Growth Large Scale Parallel Lattice Boltzmann Model of Dendritic Growth Bohumir Jelinek Mohsen Eshraghi Sergio Felicelli CAVS, Mississippi State University March 3-7, 2013 San Antonio, Texas US Army Corps of

More information

Forest-of-octrees AMR: algorithms and interfaces

Forest-of-octrees AMR: algorithms and interfaces Forest-of-octrees AMR: algorithms and interfaces Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac, Georg Stadler, Lucas C. Wilcox Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität

More information

Using the Eulerian Multiphase Model for Granular Flow

Using the Eulerian Multiphase Model for Granular Flow Tutorial 21. Using the Eulerian Multiphase Model for Granular Flow Introduction Mixing tanks are used to maintain solid particles or droplets of heavy fluids in suspension. Mixing may be required to enhance

More information

Efficiency Aspects for Advanced Fluid Finite Element Formulations

Efficiency Aspects for Advanced Fluid Finite Element Formulations Proceedings of the 5 th International Conference on Computation of Shell and Spatial Structures June 1-4, 2005 Salzburg, Austria E. Ramm, W. A. Wall, K.-U. Bletzinger, M. Bischoff (eds.) www.iassiacm2005.de

More information

High-Order Methods for Turbulent Transport in Engineering and Geosciences.

High-Order Methods for Turbulent Transport in Engineering and Geosciences. High-Order Methods for Turbulent Transport in Engineering and Geosciences. PI: Paul Fischer, University of Illinois, Urbana-Champaign, fischerp@illinois.edu Collaborators: Ananias Tomboulides, University

More information

Shape of Things to Come: Next-Gen Physics Deep Dive

Shape of Things to Come: Next-Gen Physics Deep Dive Shape of Things to Come: Next-Gen Physics Deep Dive Jean Pierre Bordes NVIDIA Corporation Free PhysX on CUDA PhysX by NVIDIA since March 2008 PhysX on CUDA available: August 2008 GPU PhysX in Games Physical

More information

Technical Report TR

Technical Report TR Technical Report TR-2015-09 Boundary condition enforcing methods for smoothed particle hydrodynamics Arman Pazouki 1, Baofang Song 2, Dan Negrut 1 1 University of Wisconsin-Madison, Madison, WI, 53706-1572,

More information

A Scalable GPU-Based Compressible Fluid Flow Solver for Unstructured Grids

A Scalable GPU-Based Compressible Fluid Flow Solver for Unstructured Grids A Scalable GPU-Based Compressible Fluid Flow Solver for Unstructured Grids Patrice Castonguay and Antony Jameson Aerospace Computing Lab, Stanford University GTC Asia, Beijing, China December 15 th, 2011

More information

Interaction of Fluid Simulation Based on PhysX Physics Engine. Huibai Wang, Jianfei Wan, Fengquan Zhang

Interaction of Fluid Simulation Based on PhysX Physics Engine. Huibai Wang, Jianfei Wan, Fengquan Zhang 4th International Conference on Sensors, Measurement and Intelligent Materials (ICSMIM 2015) Interaction of Fluid Simulation Based on PhysX Physics Engine Huibai Wang, Jianfei Wan, Fengquan Zhang College

More information