Lattice Boltzmann methods on the way to exascale
|
|
- Magdalen Walker
- 5 years ago
- Views:
Transcription
1 Lattice Boltzmann methods on the way to exascale Ulrich Rüde LSS Erlangen and CERFACS Toulouse Centre Européen de Recherche et de Formation Avancée en Calcul Scientifique Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de 1
2 Outline Motivation Building Blocks for the Direct Simulation of Complex Flows 1. Supercomputing: scalable algorithms, efficient software 2. Solid phase - rigid body dynamics 3. Fluid phase - Lattice Boltzmann method 4. Electrostatics - finite volume 5. Fast implicit solvers - multigrid 6. Gas phase - free surface tracking, volume of fluids Multi-physics applications Coupling Examples Perspectives 2
3 Motivation Simulating Additive Manufacturing Bikas, H., Stavropoulos, P., & Chryssolouris, G. (2015). Additive manufacturing methods and modelling approaches: a critical review. The International Journal of Advanced Manufacturing Technology, Klassen, A., Scharowsky, T., & Körner, C. (2014). Evaporation model for beam based additive manufacturing using free surface lattice Boltzmann methods. Journal of Physics D: Applied Physics, 47(27),
4 Electron Beam Melting Process 3D printing EU-Project Fast- EBM ARCAM (Gothenburg) TWI (Cambridge) FAU Erlangen Generation of powder bed Energy transfer by electron beam penetration depth heat transfer Flow dynamics melting melt flow surface tension wetting capillary forces contact angles solidification Ammer, R., Markl, M., Ljungblad, U., Körner, C., & UR (2014). Simulating fast electron beam melting with a parallel thermal free surface lattice Boltzmann method. Computers & Mathematics with Applications, 67(2), Ammer, R., UR, Markl, M., Jüchter V., & Körner, C. (2014). Validation experiments for LBM simulations of electron beam melting. International Journal of Modern Physics C. 4
5 SuperMuc: 3 PFlops Building Block I: Current and Future High Performance Supercomputers 5
6 Multi-PetaFlops Supercomputers Sunway TaihuLight JUQUEEN SuperMUC (phase 1) SW26010 processor 10,649,600 cores 260 cores (1.45 GHz) per node 32 GiB RAM per node 125 PFlops Peak Power consumption: MW TOP 500: #1 Blue Gene/Q architecture 458,752 PowerPC A2 cores 16 cores (1.6 GHz) per node 16 GiB RAM per node 5D torus interconnect 5.8 PFlops Peak TOP 500: #13 Intel Xeon architecture 147,456 cores 16 cores (2.7 GHz) per node 32 GiB RAM per node Pruned tree interconnect 3.2 PFlops Peak TOP 500: #27 What is the problem? LBM Methods Ulrich Rüde
7 Only optimal algorithms qualify TERRA Thought experiment with O(N 2 ) algorithm Energy computer generation gigascale: 10 9 FLOPS terascale FLOPS petascale FLOPS exascale FLOPS desired problem size DoF=N energy estimate (kwh) 1 NJoule N 2 all-to-all communication Wh 10 min of LED light 278 kwh 2 weeks blow drying hair 278 GWh 1 month electricity for Hamburg 278 PWh 100 years world electricity production TerraNeo prototype (kwh) 0.13 Wh 0.03 kwh 27 kwh? At extreme scale: optimal complexity is a must! 7
8 Building block II: The Lagrangian View: Granular media simulations with the physics engine spherical particles 256 processors time steps runtime: 48 h (including data output) texture mapping, ray tracing Pöschel, T., & Schwager, T. (2005). Computational granular dynamics: models and algorithms. Springer Science & Business Media. 8
9 Lagrangian Particle Presentation Single particle described by state variables (position x, orientation φ, translational and angular velocity v and ω) a parameterization of its shape S (e.g. geometric primitive, composite object, or mesh) and its inertia properties (mass m, principle moments of inertia I xx, I yy and I zz ). The Newton-Euler equations of motion for rigid bodies describe the rate of change of the state variables: ẋ(t) = '(t) v(t) M('(t)) =!(t) v(t) Q('(t))!(t) f(s(t),t) (s(t),t)!(t) I('(t))!(t) 9
10 Hard Contacts alternative to the discrete element method Hard contacts require impulses, exhibit non-differentiable but continuous trajectories, contact reactions are defined implicitly in general, have non-unique solutions, and can be solved numerically by methods from two classes. ) measure differential inclusions 100% 80% 60% 40% 20% 0% -20% -1 soft contacts hard contacts Fig.: Bouncing ball with a soft and a hard contact model. Moreau, J., Panagiotopoulos P. (1988): Nonsmooth mechanics and applications, vol 302. Springer, Wien-New York Popa, C., Preclik, T., & UR (2014). Regularized solution of LCP problems with application to rigid body dynamics. Numerical Algorithms, Preclik, T. & UR (2015). Ultrascale simulations of non-smooth granular dynamics; Computational Particle Mechanics, DOI: /s Extreme Scale LBM Methods - Ulrich Rüde 10
11 Newton-Euler Equations for Rigid Bodies ẋ(t) v(t) = '(t) Q('(t))!(t) v(t) f(s(t),t) M('(t)) =!(t) (s(t),t)!(t) I('(t))!(t) Contact detection minimizing signed distance function Time-continuous non-penetration constraint for hard contacts Coulomb friction Parallelization via domain partitioning Parallel data structures for contact detection Protocol for synchronization (t) 0? n (t) 0 11
12 Nonlinear Complementarity and Time Stepping Non-penetration conditions Coulomb friction conditions continuous forces =0 + =0 0? n 0 k to k 2 apple µ n + 0? n 0 k v + tok 2 to = µ n v + to + 0? n 0 k v + tok 2 to = µ n v+ to k v + to k 2 =0 impulses =0 0? n 0 k to k 2 apple µ n + 0? n 0 k v + tok 2 to = µ n v + to discrete t + v0 n( ) 0? n 0 k to k 2 apple µ n k v 0 to( )k 2 to = µ n v 0 to( ) Signorini condition impact law friction cone condition frictional reaction opposes slip Moreau, J., Panagiotopoulos P. (1988): Nonsmooth mechanics and applications, vol 302. Springer, Wien-New York Popa, C., Preclik, T., & UR (2014). Regularized solution of LCP problems with application to rigid body dynamics. Numerical Algorithms, Preclik, T. & UR (2015). Ultrascale simulations of non-smooth granular dynamics; Computational Particle Mechanics, DOI: /s
13 Regularizing Multi-Contact Problems 13
14 Parallel Computation Key features of the parallelization: domain partitioning distribution of data contact detection synchronization protocol subdomain NBGS accumulators and corrections aggressive message aggregation nearest-neighbor communication Iglberger, K., & UR (2010). Massively parallel granular flow simulations with non-spherical particles. Computer Science-Research and Development, 25(1-2), Iglberger, K., & UR (2011). Large-scale rigid body simulations. Multibody System Dynamics, 25(1),
15 Shaker scenario with sharp edged hard objects sharp-edged particles with a diameter between 0.25 mm and 2 mm. 15
16 PE marble run - rigid objects in complex geometry Animation by Sebastian Eibl and Christian Godenschwager 16
17 Scaling Results Solver algorithmically not optimal for dense systems, hence cannot scale unconditionally, but is highly efficient in many cases of practical importance Strong and weak scaling results for a constant number of iterations performed on SuperMUC and Juqueen Largest ensembles computed non-spherical particles contacts granular gas: scaling results Breakup up of compute times on Erlangen RRZE Cluster Emmy av. time per time step and 1000 particles in s av. time per time step ( rst series) av. time per time step (second series) parallel e ciency (second series) number of nodes (b) Weak-scaling graph on the Juqueen supercomputer parallel e ciency 9.5% 25.9% 8.0% 12.6% (a) Time-step profile of the granular gas executed with = 20 processes on a single node. 25.8% 18.1% 8.3% 5.9% 22.7% 30.6% 16.5% 16.0% (b) Time-step profile of the granular gas executed with = 320 processes on 16 nodes. 17
18 Building Block III: Scalable Flow Simulations with the Lattice Boltzmann Method Lallemand, P., & Luo, L. S. (2000). Theory of the lattice Boltzmann method: Dispersion, dissipation, isotropy, Galilean invariance, and stability. Physical Review E, 61(6), Feichtinger, C., Donath, S., Köstler, H., Götz, J., & Rüde, U. (2011). WaLBerla: HPC software design for computational engineering simulations. Journal of Computational Science, 2(2), Extreme Scale LBM Methods - Ulrich Rüde 18
19 The Eulerian View : Lattice-Boltzmann Discretization in squares or cubes (cells) Particles Distribution Functions (PDF) in 2D: 9 numbers (2DQ9) in 3D: D3Q19 (alternatives D3Q27, etc) Repeat (many times) stream collide Extreme Scale LBM Methods - Ulrich Rüde 19
20 Basic Lattice Boltzmann Method Single Relaxation Time (SRT) Macroscopic quantities Equilibrium distribution function Extreme Scale LBM Methods - Ulrich Rüde
21 The stream step Move PDFs into neighboring cells Non-local part, Linear propagation to neighbors (stream step) Local part, Non-linear operator, (collide step) Extreme Scale LBM Methods - Ulrich Rüde 21
22 The collide step Compute new PDFs modeling molecular collisions Most collision operators can be expressed as Equilibrium function: non-linear, depending on the conserved moments,, and. Extreme Scale LBM Methods - Ulrich Rüde 22
23 Performance on Coronary Arteries Geometry Color coded proc assignment Godenschwager, C., Schornbaum, F., Bauer, M., Köstler, H., & UR (2013). A framework for hybrid parallel flow simulations with a trillion cells in complex geometries. In Proceedings of SC13: International Conference for High Performance Computing, Networking, Storage and Analysis (p. 35). ACM. Weak scaling 458,752 cores of JUQUEEN over a trillion (10 12 ) fluid lattice cells cell sizes 1.27µm diameter of red blood cells: 7µm cell updates per second 0.41 PFlops Strong scaling 32,768 cores of SuperMUC cell sizes of 0.1 mm 2.1 million fluid cells time steps per second Extreme Scale LBM Methods - Ulrich Rüde
24 Single Node Performance JUQUEEN SuperMUC vectorized optimized standard Pohl, T., Deserno, F., Thürey, N., UR, Lammers, P., Wellein, G., & Zeiser, T. (2004). Performance evaluation of parallel largescale lattice Boltzmann applications on three supercomputing architectures. Proceedings of the 2004 ACM/IEEE conference on Supercomputing (p. 21). IEEE Computer Society. Donath, S., Iglberger, K., Wellein, G., Zeiser, T., Nitsure, A., & UR (2008). Performance comparison of different parallel lattice Boltzmann implementations on multi-core multi-socket systems. International Journal of Computational Science and Engineering, 4(1), Extreme Scale LBM - Ulrich Rüde
25 Weak scaling for TRT lid driven cavity - uniform grids JUQUEEN 16 processes per node 4 threads per process cell updates per second (TLups) SuperMUC 4 processes per node 4 threads per process cell updates per second (TLups) Körner, C., Pohl, T., UR., Thürey, N., & Zeiser, T. (2006). Parallel lattice Boltzmann methods for CFD applications. In Numerical Solution of Partial Differential Equations on Parallel Computers (pp ). Springer Berlin Heidelberg. Feichtinger, C., Habich, J., Köstler, H., UR, & Aoki, T. (2015). Performance modeling and analysis of heterogeneous lattice Boltzmann simulations on CPU GPU clusters. Parallel Computing, 46, Extreme Scale LBM - Ulrich Rüde
26 Partitioning and Parallelization static block-level refinement ( forest of octrees) static load balancing DISK compact (KiB/MiB) binary MPI IO DISK allocation of block data ( grids) separation of domain partitioning from simulation (optional) 26
27 Flow through structure of thin crystals (filter) work with Jose Pedro Galache and Antonio Gil CMT-Motores Termicos, Universitat Politecnica de Valencia 27
28 Parallel AMR load balancing different views on domain partitioning 2:1 balanced grid (used for the LBM) distributed graph: nodes = blocks edges explicitly stored as < block ID, process rank > pairs forest of octrees: octrees are not explicitly stored, but implicitly defined via block IDs 28
29 AMR and Load Balancing with walberla Isaac, T., Burstedde, C., Wilcox, L. C., & Ghattas, O. (2015). Recursive algorithms for distributed forests of octrees. SIAM Journal on Scientific Computing, 37(5), C497-C531. Meyerhenke, H., Monien, B., & Sauerwald, T. (2009). A new diffusion-based multilevel algorithm for computing graph partitions. Journal of Parallel and Distributed Computing, 69(9), Schornbaum, F., & Rüde, U. (2016). Massively Parallel Algorithms for the Lattice Boltzmann Method on NonUniform Grids. SIAM Journal on Scientific Computing, 38(2), C96-C126. Extreme Scale LBM Methods - Ulrich Rüde 29
30 AMR Performance Benchmark Environments: JUQUEEN (5.0 PFLOP/s) Blue Gene/Q, 459K cores, 1 GB/core compiler: IBM XL / IBM MPI SuperMUC (2.9 PFLOP/s) Intel Xeon, 147K cores, 2 GB/core compiler: Intel XE / IBM MPI Benchmark (LBM D3Q19 TRT): avg. blocks/process (max. blocks/proc.) level initially after refresh after load balance (1) (1) (1) (1) (9) (1) (2) (11) (4) (4) (16) (4) Extreme Scale LBM Methods - Ulrich Rüde 30
31 AMR Performance Benchmark Environments: JUQUEEN (5.0 PFLOP/s) Blue Gene/Q, 459K cores, 1 GB/core compiler: IBM XL / IBM MPI SuperMUC (2.9 PFLOP/s) Intel Xeon, 147K cores, 2 GB/core compiler: Intel XE / IBM MPI Benchmark (LBM D3Q19 TRT): during this refresh process all cells on the finest level are coarsened and coarsen refine the same amount of fine cells is created by splitting coarser cells 72 % of all cells change their size Extreme Scale LBM Methods - Ulrich Rüde 31
32 AMR Performance JUQUEEN space filling curve: Morton billion cells seconds billion cells 14 billion cells hybrid MPI+OpenMP version with SMP 1 process 2 cores 8 threads #cells per core 31, , , , ,752 cores Extreme Scale LBM Methods - Ulrich Rüde 32
33 AMR Performance JUQUEEN diffusion load balancing billion cells seconds billion cells 14 billion cells time almost independent of #processes! #cells per core 31, , , , ,752 cores Extreme Scale LBM Methods - Ulrich Rüde 33
34 Pore scale resolved flow in porous media Direct numerical simulation of flow through sphere packings Beetstra, R., Van der Hoef, M. A., & Kuipers, J. A. M. (2007). Drag force of intermediate Reynolds number flow past monoand bidisperse arrays of spheres. AIChE Journal, 53(2), Tenneti, S., Garg, R., & Subramaniam, S. (2011). Drag law for monodisperse gas solid systems using particle-resolved direct numerical simulation of flow past fixed assemblies of spheres. International journal of multiphase flow, 37(9),
35 Flow field and vorticity 2D slice visualized Domain size: Re = 300 Volume fraction:
36 Macroscopic drag correlation Finally, the drag correlation reads Average absolute percentage error: 9.7 % Bogner, S., Mohanty, S., & UR (2015). Drag correlation for dilute and moderately dense fluid-particle systems using the lattice Boltzmann method, International Journal of Multiphase Flow 68,
37 Setup of random spherical structure for porous media with the PE Massively Parallel Multiphysics Simulations - Ulrich Rüde
38 Flow over porous structure Porosity Setup Height of the channel Porosity Velocity Profile Streamwise velocity Pore geometry and streamlines Fattahi, E., Waluga, C., Wohlmuth, B., & Rüde, U. (2015). Large scale lattice Boltzmann simulation for the coupling of free and porous media flow. arxiv preprint arxiv: To appear in LNCS, Proceedings of HPCSE Massively Parallel Multiphysics Simulations - Ulrich Rüde 38
39 Turbulent flow over a permeable region Re 3000 direct numerical simulation Volume rendering of velocity magnitude Periodic in X and Y direction I10 cluster, 7x32x19=4256 core hours 1,300,000 timesteps 8 times more time steps on the finest level 39
40 Multi-Physics Simulations for Particulate Flows Parallel Coupling with walberla and PE Ladd, A. J. (1994). Numerical simulations of particulate suspensions via a discretized Boltzmann equation. Part 1. Theoretical foundation. Journal of Fluid Mechanics, 271(1), Tenneti, S., & Subramaniam, S. (2014). Particle-resolved direct numerical simulation for gas-solid flow model development. Annual Review of Fluid Mechanics, 46, Bartuschat, D., Fischermeier, E., Gustavsson, K., & UR (2016). Two computational models for simulating the tumbling motion of elongated particles in fluids. Computers & Fluids, 127,
41 Fluid-Structure Interaction direct simulation of Particle Laden Flows (4-way coupling) Götz, J., Iglberger, K., Stürmer, M., & UR (2010). Direct numerical simulation of particulate flows on processor cores. In Proceedings of Supercomputing 2010, IEEE Computer Society. Götz, J., Iglberger, K., Feichtinger, C., Donath, S., & UR (2010). Coupling multibody dynamics and computational fluid dynamics on 8192 processor cores. Parallel Computing, 36(2),
42 Simulation of sediment transport 42
43 Heterogenous CPU-GPU Simulation Particles: 31250, Domain: 400x400x200, Timesteps: Devices: 2 x M Intel Westmere, Runtime: 17.5 h Fluidized Beds: Direct numerical simulation fully resolved particles Fluid-structureinteraction 4-way-coupling C. Feichtinger, J. Habich, H. Köstler, U. Rüde, T. Aoki, Performance modeling and analysis of heterogeneous lattice Boltzmann simulations on CPU GPU clusters, Parallel Computing, Volume 46, July 2015, Pages 1-13 C. Feichtinger, J. Habich, H. Köstler, G. Hager, U. Rüde, G. Wellein, A flexible Patch-based lattice Boltzmann parallelization approach for heterogeneous GPU CPU clusters, Parallel Computing, Volume 37, Issue 9, September 2011, Pages
44 Weak Scaling experiment Götz, J., Iglberger, K., Stürmer, M., & Rüde, U. (2010, November). Direct numerical simulation of particulate flows on processor cores. In Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis (pp. 1-11). IEEE Computer Society. Efficiency Jugene Blue Gene/P Jülich Supercomputer Center 40x40x40 lattice cells per core 80x80x80 lattice cells per core Number of Cores Largest simulation on Jugene: 8 Trillion (10 12 ) variables per time step (LBM alone) Scaling 64 to cores Densely packed particles lattice cells rigid spherical objects Massively Parallel Multiphysics Simulations - Ulrich Rüde 44
45 Building Block IV (electrostatics) Direct numerical simulation of charged particles in flow Positive and negatively charged particles in flow subjected to transversal electric field Masilamani, K., Ganguly, S., Feichtinger, C., & UR (2011). Hybrid lattice-boltzmann and finite-difference simulation of electroosmotic flow in a microchannel. Fluid Dynamics Research, 43(2), Bartuschat, D., Ritter, D., & UR (2012). Parallel multigrid for electrokinetic simulation in particle-fluid flows. In High Performance Computing and Simulation (HPCS), 2012 International Conference on (pp ). IEEE. Bartuschat, D. & UR (2015). Parallel Multiphysics Simulations of Charged Particles in Microfluidic Flows, Journal of Computational Science, Volume 8, May 2015, Pages
46 6-way coupling charge distribution velocity BCs Finite volumes MG treat BCs V-cycle iter at. object motion hydrodynam. force LBM treat BCs stream-collide step electrostat. force Newtonian mechanics collision response object distance correction force Lubrication correction 46
47 Separation experiment 240 time steps fully 6-way coupled simulation 400 sec on SuperMuc weak scaling up to cores 7.1 Mio particles Total runtimes [] Number of nodes LBM Map Lubr HydrF pe MG SetRHS PtCm ElectF 10 3 MFLUPS (LBM) Number of nodes 47 LBM Perform. MG Perform MLUPS (MG)
48 Building Block V Volume of Fluids Method for Free Surface Flows joint work with Regina Ammer, Simon Bogner, Martin Bauer, Daniela Anderl, Nils Thürey, Stefan Donath, Thomas Pohl, C Körner, A. Delgado Körner, C., Thies, M., Hofmann, T., Thürey, N., & UR. (2005). Lattice Boltzmann model for free surface flow for modeling foaming. Journal of Statistical Physics, 121(1-2), Donath, S., Feichtinger, C., Pohl, T., Götz, J., & UR. (2010). A Parallel Free Surface Lattice Boltzmann Method for Large- Scale Applications. Parallel Computational Fluid Dynamics: Recent Advances and Future Directions, 318. Anderl, D., Bauer, M., Rauh, C., UR, & Delgado, A. (2014). Numerical simulation of adsorption and bubble interaction in protein foams using a lattice Boltzmann method. Food & function, 5(4), Bogner, S., Ammer, R., & Rüde, U. (2015). Boundary conditions for free interfaces with the lattice Boltzmann method. Journal of Computational Physics, 297,
49 Free Surface Flows Volume-of-Fluids like approach Flag field: Compute only in fluid Special free surface conditions in interface cells Reconstruction of curvature for surface tension 49
50 Simulation of Metal Foams Example application: Engineering: metal foam simulations Based on LBM: Free surfaces Surface tension Disjoining pressure to stabilize thin liquid films Parallelization with MPI and load balancing Other applications: Food processing Fuel cells Körner, C., Thies, M., Hofmann, T., Thürey, N., & UR (2005). Lattice Boltzmann model for free surface flow for modeling foaming. Journal of Statistical Physics, 121(1-2), Donath, S., Mecke, K., Rabha, S., Buwa, V., & UR (2011). Verification of surface tension in the parallel free surface lattice Boltzmann method in walberla. Computers & Fluids, 45(1), Thürey, N., &UR. (2009). Stable free surface flows with the lattice Boltzmann method on adaptively coarsened grids. Computing and Visualization in Science, 12(5),
51 Simulation for hygiene products (for Procter&Gamble) capillary pressure surface tension inclination contact angle 51
52 Additive Manufacturing Fast Electron Beam Melting Ammer, R., Markl, M., Ljungblad, U., Körner, C., & UR (2014). Simulating fast electron beam melting with a parallel thermal free surface lattice Boltzmann method. Computers & Mathematics with Applications, 67(2), Ammer, R., UR, Markl, M., Jüchter V., & Körner, C. (2014). Validation experiments for LBM simulations of electron beam melting. International Journal of Modern Physics C. 52
53 Simulation of Electron Beam Melting Simulating powder bed generation using the PE framework High speed camera shows melting step for manufacturing a hollow cylinder WaLBerla Simulation 53
54 Conclusions 54
55 Research in Computational Science is done by teams Harald Köstler Christian Godenschwager Kristina Pickl Regina Ammer Simon Bogner Ehsan Fattahi Florian Schornbaum Sebastian Kuckuk Christoph Rettinger Dominik Bartuschat Martin Bauer Christian Kuschel 55
56 Thank you for your attention! Videos, preprints, slides at 56
Lattice Boltzmann Methods on the way to exascale
Lattice Boltzmann Methods on the way to exascale Ulrich Rüde (LSS Erlangen, ulrich.ruede@fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de HIGH PERFORMANCE
More informationThe walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms
The walberla Framework: Multi-physics Simulations on Heterogeneous Parallel Platforms Harald Köstler, Uli Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Simulation Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationPeta-Scale Simulations with the HPC Software Framework walberla:
Peta-Scale Simulations with the HPC Software Framework walberla: Massively Parallel AMR for the Lattice Boltzmann Method SIAM PP 2016, Paris April 15, 2016 Florian Schornbaum, Christian Godenschwager,
More informationsimulation framework for piecewise regular grids
WALBERLA, an ultra-scalable multiphysics simulation framework for piecewise regular grids ParCo 2015, Edinburgh September 3rd, 2015 Christian Godenschwager, Florian Schornbaum, Martin Bauer, Harald Köstler
More informationComputational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm
Computational Fluid Dynamics with the Lattice Boltzmann Method KTH SCI, Stockholm March 17 March 21, 2014 Florian Schornbaum, Martin Bauer, Simon Bogner Chair for System Simulation Friedrich-Alexander-Universität
More informationwalberla: Developing a Massively Parallel HPC Framework
walberla: Developing a Massively Parallel HPC Framework SIAM CS&E 2013, Boston February 26, 2013 Florian Schornbaum*, Christian Godenschwager*, Martin Bauer*, Matthias Markl, Ulrich Rüde* *Chair for System
More information(LSS Erlangen, Simon Bogner, Ulrich Rüde, Thomas Pohl, Nils Thürey in collaboration with many more
Parallel Free-Surface Extension of the Lattice-Boltzmann Method A Lattice-Boltzmann Approach for Simulation of Two-Phase Flows Stefan Donath (LSS Erlangen, stefan.donath@informatik.uni-erlangen.de) Simon
More informationMassively Parallel Phase Field Simulations using HPC Framework walberla
Massively Parallel Phase Field Simulations using HPC Framework walberla SIAM CSE 2015, March 15 th 2015 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer, Harald Köstler and Ulrich
More informationSimulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method
Simulation of Liquid-Gas-Solid Flows with the Lattice Boltzmann Method June 21, 2011 Introduction Free Surface LBM Liquid-Gas-Solid Flows Parallel Computing Examples and More References Fig. Simulation
More informationPerformance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla
Performance Optimization of a Massively Parallel Phase-Field Method Using the HPC Framework walberla SIAM PP 2016, April 13 th 2016 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Johannes Hötzer,
More informationSoftware and Performance Engineering for numerical codes on GPU clusters
Software and Performance Engineering for numerical codes on GPU clusters H. Köstler International Workshop of GPU Solutions to Multiscale Problems in Science and Engineering Harbin, China 28.7.2010 2 3
More informationLarge scale Imaging on Current Many- Core Platforms
Large scale Imaging on Current Many- Core Platforms SIAM Conf. on Imaging Science 2012 May 20, 2012 Dr. Harald Köstler Chair for System Simulation Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen,
More informationReconstruction of Trees from Laser Scan Data and further Simulation Topics
Reconstruction of Trees from Laser Scan Data and further Simulation Topics Helmholtz-Research Center, Munich Daniel Ritter http://www10.informatik.uni-erlangen.de Overview 1. Introduction of the Chair
More informationPerformance and Software-Engineering Considerations for Massively Parallel Simulations
Performance and Software-Engineering Considerations for Massively Parallel Simulations Ulrich Rüde (ruede@cs.fau.de) Ben Bergen, Frank Hülsemann, Christoph Freundl Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationA Contact Angle Model for the Parallel Free Surface Lattice Boltzmann Method in walberla Stefan Donath (stefan.donath@informatik.uni-erlangen.de) Computer Science 10 (System Simulation) University of Erlangen-Nuremberg
More informationA Python extension for the massively parallel framework walberla
A Python extension for the massively parallel framework walberla PyHPC at SC 14, November 17 th 2014 Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler
More informationTowards PetaScale Computational Science
Towards PetaScale Computational Science U. Rüde (LSS Erlangen, ruede@cs.fau.de) joint work with many Lehrstuhl für Informatik 10 (Systemsimulation) Universität Erlangen-Nürnberg www10.informatik.uni-erlangen.de
More informationNumerical Algorithms on Multi-GPU Architectures
Numerical Algorithms on Multi-GPU Architectures Dr.-Ing. Harald Köstler 2 nd International Workshops on Advances in Computational Mechanics Yokohama, Japan 30.3.2010 2 3 Contents Motivation: Applications
More informationIntroducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method
Introducing a Cache-Oblivious Blocking Approach for the Lattice Boltzmann Method G. Wellein, T. Zeiser, G. Hager HPC Services Regional Computing Center A. Nitsure, K. Iglberger, U. Rüde Chair for System
More informationPerformance Analysis of the Lattice Boltzmann Method on x86-64 Architectures
Performance Analysis of the Lattice Boltzmann Method on x86-64 Architectures Jan Treibig, Simon Hausmann, Ulrich Ruede Zusammenfassung The Lattice Boltzmann method (LBM) is a well established algorithm
More informationTowards Exa-Scale: Computing with Millions of Cores
Towards Exa-Scale: Computing with Millions of Cores U. Rüde (LSS Erlangen, ruede@cs.fau.de) Lehrstuhl für Informatik 10 (Systemsimulation) Excellence Cluster Engineering of Advanced Materials Universität
More informationτ-extrapolation on 3D semi-structured finite element meshes
τ-extrapolation on 3D semi-structured finite element meshes European Multi-Grid Conference EMG 2010 Björn Gmeiner Joint work with: Tobias Gradl, Ulrich Rüde September, 2010 Contents The HHG Framework τ-extrapolation
More informationAutomatic Generation of Algorithms and Data Structures for Geometric Multigrid. Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014
Automatic Generation of Algorithms and Data Structures for Geometric Multigrid Harald Köstler, Sebastian Kuckuk Siam Parallel Processing 02/21/2014 Introduction Multigrid Goal: Solve a partial differential
More informationSimulation of moving Particles in 3D with the Lattice Boltzmann Method
Simulation of moving Particles in 3D with the Lattice Boltzmann Method, Nils Thürey, Christian Feichtinger, Hans-Joachim Schmid Chair for System Simulation University Erlangen/Nuremberg Chair for Particle
More informationDirect Numerical Simulation of Particulate Flows on Processor Cores
1 Direct Numerical Simulation of Particulate Flows on 294 912 Processor Cores Jan Götz *, Klaus Iglberger *, Markus Stürmer *, and Ulrich Rüde *,** * Chair for System Simulation, University Erlangen-Nuremberg,
More informationMassively Parallel Finite Element Simulations with deal.ii
Massively Parallel Finite Element Simulations with deal.ii Timo Heister, Texas A&M University 2012-02-16 SIAM PP2012 joint work with: Wolfgang Bangerth, Carsten Burstedde, Thomas Geenen, Martin Kronbichler
More informationHIGH PERFORMANCE COMPUTATION (HPC) FOR THE
HIGH PERFORMANCE COMPUTATION (HPC) FOR THE DEVELOPMENT OF FLUIDIZED BED TECHNOLOGIES FOR BIOMASS GASIFICATION AND CO2 CAPTURE P. Fede, H. Neau, O. Simonin Université de Toulouse; INPT, UPS ; IMFT ; 31400
More informationSustainability and Efficiency for Simulation Software in the Exascale Era
Sustainability and Efficiency for Simulation Software in the Exascale Era Dominik Thönnes, Ulrich Rüde, Nils Kohl Chair for System Simulation, University of Erlangen-Nürnberg March 09, 2018 SIAM Conference
More informationHighly efficient «on-the-fly» data processing using the open-source library CPPPO
Highly efficient «on-the-fly» data processing using the open-source library CPPPO Graz University of Technology, DCS Computing GmbH Federico Municchi, Stefan Radl, Christoph Goniva April 7 2016, Workshop
More informationLehrstuhl für Informatik 10 (Systemsimulation)
FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) On the Resource Requirements of
More informationCFD in COMSOL Multiphysics
CFD in COMSOL Multiphysics Christian Wollblad Copyright 2017 COMSOL. Any of the images, text, and equations here may be copied and modified for your own internal use. All trademarks are the property of
More informationAdaptive Hierarchical Grids with a Trillion Tetrahedra
Adaptive Hierarchical Grids with a Trillion Tetrahedra Tobias Gradl, Björn Gmeiner and U. Rüde (LSS Erlangen, ruede@cs.fau.de) in collaboration with many more Lehrstuhl für Informatik 10 (Systemsimulation)
More informationChallenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids
Challenges in Fully Generating Multigrid Solvers for the Simulation of non-newtonian Fluids Sebastian Kuckuk FAU Erlangen-Nürnberg 18.01.2016 HiStencils 2016, Prague, Czech Republic Outline Outline Scope
More informationHigh Performance Computing
High Performance Computing ADVANCED SCIENTIFIC COMPUTING Dr. Ing. Morris Riedel Adjunct Associated Professor School of Engineering and Natural Sciences, University of Iceland Research Group Leader, Juelich
More informationAdarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs)
OBJECTIVE FLUID SIMULATIONS Adarsh Krishnamurthy (cs184-bb) Bela Stepanova (cs184-bs) The basic objective of the project is the implementation of the paper Stable Fluids (Jos Stam, SIGGRAPH 99). The final
More informationGeneric finite element capabilities for forest-of-octrees AMR
Generic finite element capabilities for forest-of-octrees AMR Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität
More informationCFD MODELING FOR PNEUMATIC CONVEYING
CFD MODELING FOR PNEUMATIC CONVEYING Arvind Kumar 1, D.R. Kaushal 2, Navneet Kumar 3 1 Associate Professor YMCAUST, Faridabad 2 Associate Professor, IIT, Delhi 3 Research Scholar IIT, Delhi e-mail: arvindeem@yahoo.co.in
More informationLattice Boltzmann with CUDA
Lattice Boltzmann with CUDA Lan Shi, Li Yi & Liyuan Zhang Hauptseminar: Multicore Architectures and Programming Page 1 Outline Overview of LBM An usage of LBM Algorithm Implementation in CUDA and Optimization
More informationSupport for Multi physics in Chrono
Support for Multi physics in Chrono The Story Ahead Overview of multi physics strategy in Chrono Summary of handling rigid/flexible body dynamics using Lagrangian approach Summary of handling fluid, and
More informationSolving Partial Differential Equations on Overlapping Grids
**FULL TITLE** ASP Conference Series, Vol. **VOLUME**, **YEAR OF PUBLICATION** **NAMES OF EDITORS** Solving Partial Differential Equations on Overlapping Grids William D. Henshaw Centre for Applied Scientific
More informationLATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS
14 th European Conference on Mixing Warszawa, 10-13 September 2012 LATTICE-BOLTZMANN METHOD FOR THE SIMULATION OF LAMINAR MIXERS Felix Muggli a, Laurent Chatagny a, Jonas Lätt b a Sulzer Markets & Technology
More informationHierarchical Hybrid Grids
Hierarchical Hybrid Grids IDK Summer School 2012 Björn Gmeiner, Ulrich Rüde July, 2012 Contents Mantle convection Hierarchical Hybrid Grids Smoothers Geometric approximation Performance modeling 2 Mantle
More informationForces on particles and bubbles Literature
Lecture 1: Forces on particles and bubbles In order to describe the motion of particles and bubbles in fluids it is necessary to consider all relevant forces, external and fluid dynamic forces. Starting
More informationarxiv: v1 [cs.pf] 5 Dec 2011
Performance engineering for the Lattice Boltzmann method on GPGPUs: Architectural requirements and performance results J. Habich a, C. Feichtinger b, H. Köstler b, G. Hager a, G. Wellein a,b a Erlangen
More informationDevelopment of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics
Development of an Integrated Computational Simulation Method for Fluid Driven Structure Movement and Acoustics I. Pantle Fachgebiet Strömungsmaschinen Karlsruher Institut für Technologie KIT Motivation
More informationAlgorithms, System and Data Centre Optimisation for Energy Efficient HPC
2015-09-14 Algorithms, System and Data Centre Optimisation for Energy Efficient HPC Vincent Heuveline URZ Computing Centre of Heidelberg University EMCL Engineering Mathematics and Computing Lab 1 Energy
More informationDevelopment of Hybrid Fluid Jet / Float Polishing Process
COMSOL Conference - Tokyo 2013 Development of Hybrid Fluid Jet / Float Polishing Process A. Beaucamp, Y. Namba Dept. of Mechanical Engineering, Chubu University, Japan Zeeko LTD, United Kingdom Research
More informationFinal drive lubrication modeling
Final drive lubrication modeling E. Avdeev a,b 1, V. Ovchinnikov b a Samara University, b Laduga Automotive Engineering Abstract. In this paper we describe the method, which is the composition of finite
More informationApplications of ICFD /SPH Solvers by LS-DYNA to Solve Water Splashing Impact to Automobile Body. Abstract
Applications of ICFD /SPH Solvers by LS-DYNA to Solve Water Splashing Impact to Automobile Body George Wang (1 ), Kevin Gardner (3), Eric DeHoff (1), Facundo del Pin (2), Inaki Caldichoury (2), Edouard
More informationFree Surface Flows with Moving and Deforming Objects for LBM
Free Surface Flows with Moving and Deforming Objects for LBM Nils Thürey, Klaus Iglberger, Ulrich Rüde Institute for System Simulation (LSS), University of Erlangen-Nuremberg, Cauerstr. 6, 91058 Erlangen,
More informationPerformance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures
Performance and Accuracy of Lattice-Boltzmann Kernels on Multi- and Manycore Architectures Dirk Ribbrock, Markus Geveler, Dominik Göddeke, Stefan Turek Angewandte Mathematik, Technische Universität Dortmund
More informationFree Surface Lattice-Boltzmann fluid simulations. with and without level sets.
ree Surface Lattice-Boltzmann fluid simulations with and without level sets Nils Thürey, Ulrich Rüde University of Erlangen-Nuremberg System Simulation roup Cauerstr. 6, 91054 Erlangen, ermany Email: Nils.Thuerey@cs.fau.de
More informationPreliminary Spray Cooling Simulations Using a Full-Cone Water Spray
39th Dayton-Cincinnati Aerospace Sciences Symposium Preliminary Spray Cooling Simulations Using a Full-Cone Water Spray Murat Dinc Prof. Donald D. Gray (advisor), Prof. John M. Kuhlman, Nicholas L. Hillen,
More informationTowards Generating Solvers for the Simulation of non-newtonian Fluids. Harald Köstler, Sebastian Kuckuk FAU Erlangen-Nürnberg
Towards Generating Solvers for the Simulation of non-newtonian Fluids Harald Köstler, Sebastian Kuckuk FAU Erlangen-Nürnberg 22.12.2015 Outline Outline Scope and Motivation Project ExaStencils The Application
More informationPerformance Metrics of a Parallel Three Dimensional Two-Phase DSMC Method for Particle-Laden Flows
Performance Metrics of a Parallel Three Dimensional Two-Phase DSMC Method for Particle-Laden Flows Benzi John* and M. Damodaran** Division of Thermal and Fluids Engineering, School of Mechanical and Aerospace
More informationParticleworks: Particle-based CAE Software fully ported to GPU
Particleworks: Particle-based CAE Software fully ported to GPU Introduction PrometechVideo_v3.2.3.wmv 3.5 min. Particleworks Why the particle method? Existing methods FEM, FVM, FLIP, Fluid calculation
More informationPossibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics
Possibility of Implicit LES for Two-Dimensional Incompressible Lid-Driven Cavity Flow Based on COMSOL Multiphysics Masanori Hashiguchi 1 1 Keisoku Engineering System Co., Ltd. 1-9-5 Uchikanda, Chiyoda-ku,
More informationEnzo-P / Cello. Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology. San Diego Supercomputer Center. Department of Physics and Astronomy
Enzo-P / Cello Scalable Adaptive Mesh Refinement for Astrophysics and Cosmology James Bordner 1 Michael L. Norman 1 Brian O Shea 2 1 University of California, San Diego San Diego Supercomputer Center 2
More informationORAP Forum October 10, 2013
Towards Petaflop simulations of core collapse supernovae ORAP Forum October 10, 2013 Andreas Marek 1 together with Markus Rampp 1, Florian Hanke 2, and Thomas Janka 2 1 Rechenzentrum der Max-Planck-Gesellschaft
More informationOptimizing Bio-Inspired Flow Channel Design on Bipolar Plates of PEM Fuel Cells
Excerpt from the Proceedings of the COMSOL Conference 2010 Boston Optimizing Bio-Inspired Flow Channel Design on Bipolar Plates of PEM Fuel Cells James A. Peitzmeier *1, Steven Kapturowski 2 and Xia Wang
More informationDirect Numerical Simulation of Turbulent Boundary Layers at High Reynolds Numbers.
Direct Numerical Simulation of Turbulent Boundary Layers at High Reynolds Numbers. G. Borrell, J.A. Sillero and J. Jiménez, Corresponding author: guillem@torroja.dmt.upm.es School of Aeronautics, Universidad
More informationCGT 581 G Fluids. Overview. Some terms. Some terms
CGT 581 G Fluids Bedřich Beneš, Ph.D. Purdue University Department of Computer Graphics Technology Overview Some terms Incompressible Navier-Stokes Boundary conditions Lagrange vs. Euler Eulerian approaches
More informationSIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC PARTICLE CONFIGURATIONS
XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22, 2012 SIMULATION OF FLOW FIELD AROUND AND INSIDE SCOUR PROTECTION WITH PHYSICAL AND REALISTIC
More informationLattice Boltzmann Liquid Simulation with Moving Objects on Graphics Hardware
Lattice Boltzmann Liquid Simulation with Moving Objects on Graphics Hardware Duncan Clough Supervised by: James Gain, Michelle Kuttel 1 Introduction Fluid simulation is a common element in recent feature
More informationLARGE-SCALE FREE-SURFACE FLOW SIMULATION USING LATTICE BOLTZMANN METHOD ON MULTI-GPU CLUSTERS
ECCOMAS Congress 2016 VII European Congress on Computational Methods in Applied Sciences and Engineering M. Papadrakakis, V. Papadopoulos, G. Stefanou, V. Plevris (eds.) Crete Island, Greece, 5 10 June
More informationA Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla
A Python Extension for the Massively Parallel Multiphysics Simulation Framework walberla Martin Bauer, Florian Schornbaum, Christian Godenschwager, Matthias Markl, Daniela Anderl, Harald Köstler, and Ulrich
More informationHigh performance computing and numerical modeling
High performance computing and numerical modeling Volker Springel Plan for my lectures Lecture 1: Collisional and collisionless N-body dynamics Lecture 2: Gravitational force calculation Lecture 3: Basic
More informationENERGY-224 Reservoir Simulation Project Report. Ala Alzayer
ENERGY-224 Reservoir Simulation Project Report Ala Alzayer Autumn Quarter December 3, 2014 Contents 1 Objective 2 2 Governing Equations 2 3 Methodolgy 3 3.1 BlockMesh.........................................
More informationFRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)
FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) walberla: Visualization of Fluid
More informationFOUR WHAT S NEW IN THIS VERSION? 4.1 FLOW-3D Usability CHAPTER
CHAPTER FOUR WHAT S NEW IN THIS VERSION? FLOW-3D v11.2.0 continues to streamline engineers simulation workflows by enabling them to more quickly set up simulations, avoid common errors, identify and enter
More informationIntroduction to C omputational F luid Dynamics. D. Murrin
Introduction to C omputational F luid Dynamics D. Murrin Computational fluid dynamics (CFD) is the science of predicting fluid flow, heat transfer, mass transfer, chemical reactions, and related phenomena
More informationDynamics in Maya. Gary Monheit Alias Wavefront PHYSICALLY BASED MODELING SH1 SIGGRAPH 97 COURSE NOTES
Dynamics in Maya Gary Monheit Alias Wavefront SH1 Dynamics in Maya Overall Requirements Architecture and Features Animations SH2 Overall Requirements Why Dynamics? Problems with traditional animation techniques
More informationA Peta-scale LES (Large-Eddy Simulation) for Turbulent Flows Based on Lattice Boltzmann Method
GTC (GPU Technology Conference) 2013, San Jose, 2013, March 20 A Peta-scale LES (Large-Eddy Simulation) for Turbulent Flows Based on Lattice Boltzmann Method Takayuki Aoki Global Scientific Information
More informationDriven Cavity Example
BMAppendixI.qxd 11/14/12 6:55 PM Page I-1 I CFD Driven Cavity Example I.1 Problem One of the classic benchmarks in CFD is the driven cavity problem. Consider steady, incompressible, viscous flow in a square
More informationPhys 113 Final Project
Phys 113 Final Project LBM model for capillary fluid modelling Andrey Sushko Introduction Lattice Boltzmann Method (LMB) models have recently emerged as a highly effective means of modelling complex fluid
More informationParallel High-Order Geometric Multigrid Methods on Adaptive Meshes for Highly Heterogeneous Nonlinear Stokes Flow Simulations of Earth s Mantle
ICES Student Forum The University of Texas at Austin, USA November 4, 204 Parallel High-Order Geometric Multigrid Methods on Adaptive Meshes for Highly Heterogeneous Nonlinear Stokes Flow Simulations of
More informationAbstract. 1 Introduction
On the Use of Pore-Scale Computational Models for Two-Phase Porous-media Flows M.A. Celia, P.C Reeves, H.K. Dahle Environmental Engineering and Water Resources Program, Department of Civil Engineering
More informationLS-DYNA 980 : Recent Developments, Application Areas and Validation Process of the Incompressible fluid solver (ICFD) in LS-DYNA.
12 th International LS-DYNA Users Conference FSI/ALE(1) LS-DYNA 980 : Recent Developments, Application Areas and Validation Process of the Incompressible fluid solver (ICFD) in LS-DYNA Part 1 Facundo Del
More informationHigh Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters
SIAM PP 2014 High Scalability of Lattice Boltzmann Simulations with Turbulence Models using Heterogeneous Clusters C. Riesinger, A. Bakhtiari, M. Schreiber Technische Universität München February 20, 2014
More informationI/O at JSC. I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O. Wolfgang Frings
Mitglied der Helmholtz-Gemeinschaft I/O at JSC I/O Infrastructure Workloads, Use Case I/O System Usage and Performance SIONlib: Task-Local I/O Wolfgang Frings W.Frings@fz-juelich.de Jülich Supercomputing
More informationNIA CFD Futures Conference Hampton, VA; August 2012
Petascale Computing and Similarity Scaling in Turbulence P. K. Yeung Schools of AE, CSE, ME Georgia Tech pk.yeung@ae.gatech.edu NIA CFD Futures Conference Hampton, VA; August 2012 10 2 10 1 10 4 10 5 Supported
More informationSimulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg
Simulation in Computer Graphics Deformable Objects Matthias Teschner Computer Science Department University of Freiburg Outline introduction forces performance collision handling visualization University
More informationSPH: Why and what for?
SPH: Why and what for? 4 th SPHERIC training day David Le Touzé, Fluid Mechanics Laboratory, Ecole Centrale de Nantes / CNRS SPH What for and why? How it works? Why not for everything? Duality of SPH SPH
More informationRealtime Water Simulation on GPU. Nuttapong Chentanez NVIDIA Research
1 Realtime Water Simulation on GPU Nuttapong Chentanez NVIDIA Research 2 3 Overview Approaches to realtime water simulation Hybrid shallow water solver + particles Hybrid 3D tall cell water solver + particles
More informationParticle Tracing Module
Particle Tracing Module Particle Tracing Module Released with version 4.2a in October 2011 Add-on to COMSOL Multiphysics Combines with any COMSOL Multiphysics Module Particle Tracing Particle tracing
More informationEfficient Imaging Algorithms on Many-Core Platforms
Efficient Imaging Algorithms on Many-Core Platforms H. Köstler Dagstuhl, 22.11.2011 Contents Imaging Applications HDR Compression performance of PDE-based models Image Denoising performance of patch-based
More informationPart I: Theoretical Background and Integration-Based Methods
Large Vector Field Visualization: Theory and Practice Part I: Theoretical Background and Integration-Based Methods Christoph Garth Overview Foundations Time-Varying Vector Fields Numerical Integration
More informationCoupling of STAR-CCM+ to Other Theoretical or Numerical Solutions. Milovan Perić
Coupling of STAR-CCM+ to Other Theoretical or Numerical Solutions Milovan Perić Contents The need to couple STAR-CCM+ with other theoretical or numerical solutions Coupling approaches: surface and volume
More informationMESHLESS SOLUTION OF INCOMPRESSIBLE FLOW OVER BACKWARD-FACING STEP
Vol. 12, Issue 1/2016, 63-68 DOI: 10.1515/cee-2016-0009 MESHLESS SOLUTION OF INCOMPRESSIBLE FLOW OVER BACKWARD-FACING STEP Juraj MUŽÍK 1,* 1 Department of Geotechnics, Faculty of Civil Engineering, University
More informationDirect Numerical Simulation of a Low Pressure Turbine Cascade. Christoph Müller
Low Pressure NOFUN 2015, Braunschweig, Overview PostProcessing Experimental test facility Grid generation Inflow turbulence Conclusion and slide 2 / 16 Project Scale resolving Simulations give insight
More informationComputational Fluid Dynamic Hydraulic Characterization: G3 Cube vs. Dolos Armour Unit. IS le Roux, WJS van der Merwe & CL de Wet
Computational Fluid Dynamic Hydraulic Characterization: G3 Cube vs. Dolos Armour Unit IS le Roux, WJS van der Merwe & CL de Wet Presentation Outline Scope. Assumptions and boundary values. Numerical mesh.
More informationLarge Scale Parallel Lattice Boltzmann Model of Dendritic Growth
Large Scale Parallel Lattice Boltzmann Model of Dendritic Growth Bohumir Jelinek Mohsen Eshraghi Sergio Felicelli CAVS, Mississippi State University March 3-7, 2013 San Antonio, Texas US Army Corps of
More informationForest-of-octrees AMR: algorithms and interfaces
Forest-of-octrees AMR: algorithms and interfaces Carsten Burstedde joint work with Omar Ghattas, Tobin Isaac, Georg Stadler, Lucas C. Wilcox Institut für Numerische Simulation (INS) Rheinische Friedrich-Wilhelms-Universität
More informationUsing the Eulerian Multiphase Model for Granular Flow
Tutorial 21. Using the Eulerian Multiphase Model for Granular Flow Introduction Mixing tanks are used to maintain solid particles or droplets of heavy fluids in suspension. Mixing may be required to enhance
More informationEfficiency Aspects for Advanced Fluid Finite Element Formulations
Proceedings of the 5 th International Conference on Computation of Shell and Spatial Structures June 1-4, 2005 Salzburg, Austria E. Ramm, W. A. Wall, K.-U. Bletzinger, M. Bischoff (eds.) www.iassiacm2005.de
More informationHigh-Order Methods for Turbulent Transport in Engineering and Geosciences.
High-Order Methods for Turbulent Transport in Engineering and Geosciences. PI: Paul Fischer, University of Illinois, Urbana-Champaign, fischerp@illinois.edu Collaborators: Ananias Tomboulides, University
More informationShape of Things to Come: Next-Gen Physics Deep Dive
Shape of Things to Come: Next-Gen Physics Deep Dive Jean Pierre Bordes NVIDIA Corporation Free PhysX on CUDA PhysX by NVIDIA since March 2008 PhysX on CUDA available: August 2008 GPU PhysX in Games Physical
More informationTechnical Report TR
Technical Report TR-2015-09 Boundary condition enforcing methods for smoothed particle hydrodynamics Arman Pazouki 1, Baofang Song 2, Dan Negrut 1 1 University of Wisconsin-Madison, Madison, WI, 53706-1572,
More informationA Scalable GPU-Based Compressible Fluid Flow Solver for Unstructured Grids
A Scalable GPU-Based Compressible Fluid Flow Solver for Unstructured Grids Patrice Castonguay and Antony Jameson Aerospace Computing Lab, Stanford University GTC Asia, Beijing, China December 15 th, 2011
More informationInteraction of Fluid Simulation Based on PhysX Physics Engine. Huibai Wang, Jianfei Wan, Fengquan Zhang
4th International Conference on Sensors, Measurement and Intelligent Materials (ICSMIM 2015) Interaction of Fluid Simulation Based on PhysX Physics Engine Huibai Wang, Jianfei Wan, Fengquan Zhang College
More information