Prediction of Parallel NIKE3D Performance on the KSR1 System. P. s. su R. E. Fulton T. Zacharia

Similar documents
Alignment and Micro-Inspection System

FY97 ICCS Prototype Specification

Electronic Weight-and-Dimensional-Data Entry in a Computer Database

Introduction to Abaqus. About this Course

Visualization for the Large Scale Data Analysis Project. R. E. Flanery, Jr. J. M. Donato

A comparison of DYNA3D, NIKE3D and LS-DYNA

7 h e submitted manuscript has been authored by a contractor of the U.S. Government under contract No. DEAccordingty. the U.S.

MAS. &lliedsignal. Design of Intertransition Digitizing Decomutator KCP Federal Manufacturing & Technologies. K. L.

Adding a System Call to Plan 9

GA A22637 REAL TIME EQUILIBRIUM RECONSTRUCTION FOR CONTROL OF THE DISCHARGE IN THE DIII D TOKAMAK

LosAlamos National Laboratory LosAlamos New Mexico HEXAHEDRON, WEDGE, TETRAHEDRON, AND PYRAMID DIFFUSION OPERATOR DISCRETIZATION

REAL-TIME MULTIPLE NETWORKED VIEWER CAPABILITY OF THE DIII D EC DATA ACQUISITION SYSTEM

METADATA REGISTRY, ISO/IEC 11179

NIF ICCS Test Controller for Automated & Manual Testing

Parallel Execution of Functional Mock-up Units in Buildings Modeling

ENDF/B-VII.1 versus ENDFB/-VII.0: What s Different?

Final Report for LDRD Project Learning Efficient Hypermedia N avi g a ti o n

High Scalability Resource Management with SLURM Supercomputing 2008 November 2008

OPTIMIZING CHEMICAL SENSOR ARRAY SIZES

Example 24 Spring-back

Testing of PVODE, a Parallel ODE Solver

Portable Data Acquisition System

ESNET Requirements for Physics Reseirch at the SSCL

Thomas Zacharia Metals and Ceramics Oak Ridge National Laboratory. Oak Ridge, TN (615) (FAX: )

In-Field Programming of Smart Meter and Meter Firmware Upgrade

Solid and shell elements

CHAPTER 1. Introduction

SOLIDWORKS. SOLIDWORKS Simulation Premium: Nonlinear. Dassault Systèmes SolidWorks Corporation 175 Wyman Street Waltham, Massachusetts USA

ADAPTIVE MESH REFINEMENT IN CTH

SIMULATION TOOLS FOR HAZARDOUS WASTE REMOVAL*

ACCELERATOR OPERATION MANAGEMENT USING OBJECTS*

Go SOLAR Online Permitting System A Guide for Applicants November 2012

Implementation of the AES as a Hash Function for Confirming the Identity of Software on a Computer System

Large Scale Test Simulations using the Virtual Environment for Test Optimization

Java Based Open Architecture Controller

ALAMO: Automatic Learning of Algebraic Models for Optimization

Stereo Vision Based Automated Grasp Planning

Clusters Using Nonlinear Magnification

SmartSacramento Distribution Automation

Testing PL/SQL with Ounit UCRL-PRES

A general contact detection algorithm for finite element analysis

A METHOD FOR EFFICIENT FRACTIONAL SAMPLE DELAY GENERATION FOR REAL-TIME FREQUENCY-DOMAIN BEAMFORMERS

Reduced Order Models for Oxycombustion Boiler Optimization

Engineers can be significantly more productive when ANSYS Mechanical runs on CPUs with a high core count. Executive Summary

Modeling Strategies for Dynamic Finite Element Cask Analyses

Mike McGlaun, Allen Robinson and James Peery. Sandia National Laboratories, Albuquerque, New Mexico, U.S.A.

Structural Nonlinear Analysis of a Disconnector Pelikelo (a kind of hook) using ANSYS/Workbench

Embedded Reinforcements

Graphical Programming of Telerobotic Tasks

DERIVATIVE-FREE OPTIMIZATION ENHANCED-SURROGATE MODEL DEVELOPMENT FOR OPTIMIZATION. Alison Cozad, Nick Sahinidis, David Miller

and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof.

A Multiple Constraint Approach for Finite Element Analysis of Moment Frames with Radius-cut RBS Connections

On Demand Meter Reading from CIS

COMPUTATIONAL FLUID DYNAMICS (CFD) ANALYSIS AND DEVELOPMENT OF HALON- REPLACEMENT FIRE EXTINGUISHING SYSTEMS (PHASE II)

MOLECULAR DYNAMICS ON DISTRIBUTED-MEMORY MIMD COMPUTERS WITH LOAD BALANCING 1. INTRODUCTION

Centrally Managed. Ding Jun JLAB-ACC This process of name resolution and control point location

Modelling Flat Spring Performance Using FEA

Introduction to Finite Element Analysis using ANSYS

Bridging The Gap Between Industry And Academia

Module 1: Introduction to Finite Element Analysis. Lecture 4: Steps in Finite Element Analysis

CHAPTER 6 EXPERIMENTAL AND FINITE ELEMENT SIMULATION STUDIES OF SUPERPLASTIC BOX FORMING

Information to Insight

Development of Web Applications for Savannah River Site

MSC.Patran Marc Preference

studying of the prying action effect in steel connection

Analysis of Fluid-Structure Interaction Effects of Liquid-Filled Container under Drop Testing

Abaqus Technology Brief. Two-Pass Rolling Simulation

Recent Advances on Higher Order 27-node Hexahedral Element in LS-DYNA

Optimizing Bandwidth Utilization in Packet Based Telemetry Systems. Jeffrey R Kalibjian

Charles L. Bennett, USA. Livermore, CA 94550

SAMCEF for ROTORS. Chapter 3.2: Rotor modeling. This document is the property of SAMTECH S.A. MEF A, Page 1

for use with the WINDOW 4.1 Computer Program

Parallel finite element algorithms for the analysis of multiscale plasticity problems

Print Depth Prediction in Hot Forming Process with a Reconfigurable Die

Using MSC.Nastran for Explicit FEM Simulations

Analysis of fluid-solid coupling vibration characteristics of probe based on ANSYS Workbench

Computer Life (CPL) ISSN: Finite Element Analysis of Bearing Box on SolidWorks

Los Alamos National Laboratory

Multi-scale Material Modeling Applied from Specimen to Full Car Level using LS-DYNA

Some Aspects for the Simulation of a Non-Linear Problem with Plasticity and Contact

zorder-lib: Library API for Z-Order Memory Layout

VALIDATE SIMULATION TECHNIQUES OF A MOBILE EXPLOSIVE CONTAINMENT VESSEL

Using the Workbench LS-DYNA Extension

Efficient Topology, Topometry and Sizing Optimisation for LS-DYNA Analysis Problems. Coupling LS-DYNA to VR&D GENESIS

The NMT-5 Criticality Database

NFRC Spectral Data Library #4 for use with the WINDOW 4.1 Computer Program

Efficient Shape Optimisation of an Aircraft Landing Gear Door Locking Mechanism by Coupling Abaqus to GENESIS

CHAPTER 7 BEHAVIOUR OF THE COMBINED EFFECT OF ROOFING ELEMENTS

Finite element method - tutorial no. 1

Vehicle Load Area Division Wall Integrity during Frontal Crash

Intelligent Grid and Lessons Learned. April 26, 2011 SWEDE Conference

SIMULATION CAPABILITIES IN CREO

Contributors: Surabhi Jain, Gengbin Zheng, Maria Garzaran, Jim Cownie, Taru Doodi, and Terry L. Wilmarth

PJM Interconnection Smart Grid Investment Grant Update

Early Experiences with Trinity - The First Advanced Technology Platform for the ASC Program

Chapter 5 Modeling and Simulation of Mechanism

Collaborating with Human Factors when Designing an Electronic Textbook

SIMULATION CAPABILITIES IN CREO. Enhance Your Product Design with Simulation & Analysis

MULTIPLE HIGH VOLTAGE MODULATORS OPERATING INDEPENDENTLY FROM A SINGLE COMMON 100 kv dc POWER SUPPLY

R. P. Evans (INEEL) J. G. Keller (INEEL) A. G. Stephens (ISU) J. Blotter (ISU) May 17, 1999 May 20, 1999

Transcription:

Prediction of Parallel NIKE3D Performance on the KSR1 System P. s. su R. E. Fulton T. Zacharia

This report has been reproduced directly from the best available copy. Available to DOE and DOE contractors from the Office of Scientific and Technical Information, P.0. Box 62, Oak Ridge, TN 37831; prices available from (615) 576-840 1, FTS 626-840 1. Available to the public from the National Technical Information Service, US. Department of Commerce. 5285 Port Royal Rd., Springfield, VA 22161. c This report was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor any agency thereof, nor any of their employees, makes any warranty, express or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States Government or any agency thereof. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof..

DISCLAIMER Portions of this document may be illegible in electronic image products. Images are produced from the best available original document.

ORNLm-12732 Metals and Ceramics Division PREDICI'ION OF PARALLEL NIRE3D PERFORMANCE ON THE KSRl SYSTEM T. Zacharia, P. S. Su, and R. E. Fulton Date Published May 1995 NOTICE: This document contains information of a preliminary nature. It is subject to revision or correction and therefore does not represent a final report. Prepared for U.S.Department of Energy Office of Basic Energy Sciences KC 02 01 05 0 Prepared by the OAK RIDGE NATIONAL LABORATORY Oak Ridge, Tennessee 3783 1-6285 managed by MARTIN MARIETI+AENERGY SYSTEMS, INC. for the U.S. DEPARTMENT OF ENERGY

.

LISTOFF'IGURES... V... vii ABSTRACT... 1 1. INTRODUCTION... 1.1 NIKE3D FINITE ELEMENT CODE... 1.2 KSRl MULTIPROCESSOR SYSTEM... 1 2 3 2. NCJMERICALEXAMPLES... 2.1 PIPE WHIP TEST PROBLEM... 2.2 BAR IMPACTING RIGID WALL TEST PROBLEM... 3 5 5... 5 9 9 9 10 10 LISTOFTABLES 3. PREDICTION OF PARALLEL NIKE3D PERFORMANCE 3.1 3.2 3.3 3.4 3.5 PARALLEL MATRIX DECOMPOSITION... PARALLEL ELEMENT MATRIX GENERATION... PARALLEX GLOBAL STIFFNESS MATRIX ASSEMBLY... PARALLEL FORWARDBACKWARD SUBSTITUTION............. PREDICTION OF PARALLEL NIKE3D PERFORMANCE... CONCLUSIONS.... 12 5. REFERENCES... 12 4.

LIST OF FIGURES Figure 1 The architecture of the KSRl multiprocessor system 2 The pipe whip test problem 3 The bar impacting wall test problem......... V 4 6 8

Table 1 Central processing unit (CPU) time for each major step in the serialversion... 7 2 Predicted parallel performance for the pipe whip test problem on theksrlsystem... 11 3 Predicted parallel performance for the bar impacting wall test problem on the KSRl system... 11 vii

I PREDICI'ION OF PARALLEL NIRE3D PERFORMANCE ON THE KSRl SYSTEM* Philip S. Sut, Robert E. Fulton*, and Thomas Zacharia r Finite element method is one of the bases for numerical solutions to engineering problems. Complex engineering problems using finite element analysis typically imply excessively large computational time. Parallel supercomputers have the potential for significantly increasing calculation speeds in order to meet these computational requirements. This paper predicts parallel NIJE3D performance on the Kendall Square Research (KSR1) system. The first part of the prediction is based on the implementation of parallel Cholesky (UTDU) matrix decomposition algorithm through actual computations on the KSRl multiprocessor system, with 64 processors, at Oak Ridge National Laboratory. The other predictions are based on actual computations for parallel element matrix generation, parallel global stiffness matrix assembly, and parallel forwardjbachard substitution on the BBN TC2000 multiprocessor system at Lawrence Livermore National Laboratory. The preliminary results indicate that parallel NIKE3D performance can be attractive under localhhared-memory multiprocessor system environments. Finite element method is one of the bases for numerical solutions to engineering problems. Solving the associated simultaneous equations can take up to 50% of the total computation time. Realistic engineering problems using such analysis typically require excessively large amounts of computation time. However, the emergence of parallel computers has revolutionized approaches to numerical solutions of large-scale engineering problems. New algorithms are being developed by researchers in engineering and computer science in order to exploit the tremendous potential of parallel computing. Parallel computers * This research was supported in part by an appointment to the Oak Ridge National Laboratory Postdoctoral Research Associates Program administered jointly by the Oak Ridge National Laboratory and the Oak Ridge Institute for Science and Education and was sponsored by the Division of Materials Sciences, U.S. Department of Energy, under contract DE-AC05-840R21400 with Martin Marietta Energy Systems, Inc. tpostdoctoral Research Associate, Oak Ridge National Laboratory, Oak Ridge, TN 37831-6140. *Professor, School of Mechanical Engineering, Georgia Institute of Technology, Atlanta, GA 30332-0405. 1

2 have become important tools for solving complex and computationally intensive science and engineering problems. Parallel supercomputers have the potential for significantly increasing calculation speeds in order to meet these computational requirements. This paper predicts parallel NIKE3D performance on the KendalI Square Research (KSR1) system. The first part of the prediction is based on the implementation of parallel Cholesky (UTDU) matrix decomposition algorithm through actual computations on the KSRl multiprocessor system, with 64 processors, at Oak Ridge National Laboratory (ORNL).' The other predictions are based on actual computations for parallel element matrix generation, parallel global stiffness matrix assembly, and parallel forwardbackward substitution on the BBN TC2000 multiprocessor system at Lawrence Livermore National Laboratory (LLNL).2 The preliminary results indicate that parallel NIKE3D performance can be attractive under local/shared-memory multiprocessor system environments. 1.1 NIRE3DFINITEELEMENTCODE -3D (ref. 3) is an implicit finite element code for analyzing the finite strain static and dynamic response of three-dimensional (3-D) inelastic solids, shells, and beams. The finite element formulation accounts for both material and geometric nonlinearities. The code is fully vectorized and available on several computer platforms. A number of material models are incorporated to simulate a wide range of material behavior, including elastoplasticity, anisotropy, creep, thermal effects, and rate dependence. Slideline algorithms model gaps and sliding along material interfaces, including interface friction and single surface contact. Interactive graphics and rezoning are available for analyses with large mesh distortions. Several nonlinear solution strategies are available, including full-, modified-, and quasi-newton methods. The resulting system of simultaneous linear equations is either solved iteratively by an element-by-element method or directly by a factorization method, for which case bandwidth minimization is optional. Data may be stored either in or out of core memory to allow for large analysis. As a Lagrangian implicit code, NIKE3D is ideally suited for quasi-static and low-rate dynamic problems of solids, shells, and beams. NIKE3D uses a small number of relatively large time steps. A coupled system of nonlinear algebraic equations is solved at each time step by a linearization and iteration procedure. A robust and efficient nonlinear solution strategy is an essential capability for the severely nonlinear problems typically solved by NIKE3D. I

3 Ongoing parallel computation research and development activities at ORNL have focused on exploring parallelization of NIKE3D's application on locahhared-memory multiprocessor systems such as the KSR1. Since the equation solver can use up to one-half of the total computation time, parallelization of the popular direct matrix factorization algorithm on the multiprocessor system for nonlinear problems has been tested.' The parallel Cholesky (UTDU) matrix decomposition algorithm and skyline data storage scheme were used for the implementation. 1 2 KSRl MULTIPRO(SESS0R!iX?IEM The KSRl architecture is a multiprocessor system composed of a hierarchy of rings. The lowest level, ring:o, consists of a 34-slot backplane connecting 32 processing cells and 2 cells responsible for routing to the next higher layer ring, ring:l. A fully populated ring:l is composed of the interconnecting cells from 32 ring:o rings. A fully configured KSRl multiprocessor system is composed of two layers containing 1024 processing cells along with two ring interconnecting cells on each ring:o. Figure 1 shows the hierarchical ring structure of the KSRl multiprocessor. Each processor contains a 256-KB data cache and a 256-KB instruction cache. The on-board data and instruction caches are referred to as mbcaches. A daughter board connected to each processing cell contains 32 MB of memory referred to as local cache. The word size of the KSRl multiprocessor system is 64 bits, and all functional units are based on 64-bit operands. All of the local caches make up the ALLCACHE memory system. The KSRl architecture and chip set are designed specifically to support a locallshared-memory multiprocessor. Reference 4 provides more detail on the actual implementation. Each processor has the peak performance of 40 MFLOPS. The KSRl multiprocessor system at ORNL has 64 (2 ring:o rings) processors. 2 NUMERICALEXAMPLES Serial and parallel NIKE3D were implemented for two nonlinear, dynamic, two- dimensional, and 3-D finite element problems. There are four major steps in NIKE3D: matrix decomposition, element matrix generation, global stiffness matrix assembly, and forward/backward substitution. The NIKE3D applications were run in core mode to eliminate the disk input/output timing. The prediction is based on computations that were performed

4 ORNL-DWG 95-5858 Ring:O Ring:O Ring:O Ring:O Ring:O 0 32 processor cells Ring:l Up to 32 Ring:Os Up to 1024 processor cells Fig. 1. The architecture of the KSRl multiprocessor system.

5 either on the KSRl multiprocessor system, with 64 processors, at ORNL or the BBN TC2000 multiprocessor system, with 104 processors, at LLNL. 2 1 PIPE WHIP TEST PROBLEM This transient dynamic analysis simulates the impact of two steel pipes? The pipes have 3.3125-in. OD, 0.432411. wall thickness, and are 50 in. long. The target pipe is supported with a fixed boundary condition at each end. The second pipe swings freely about one end with an angular velocity at impact of 50 rad/s. One-half of this symmetric problem is modeled. Shell elements are used in the model, and a slide surface (type 3) is included between the pipes. There are 1484 shell elements, 8985 D. E, and 20 time steps are used with a fixed step size of 5.0 s. All subroutines were compiled with the KSRl optimization option up to the first x and second levels. The central processing unit (CPU)time for the serial version is 2362 s. Figure 2 (ref. 3) shows the deformed geometry of the pipe whip test problem at several stages of analysis. The CPU times for each of the major steps and their percentages are shown in Table 1. 2 2 BAR IMPACTING RIGID WALL TEST PROBLEM This transient dynamic analysis simulates a copper bar of 6.4 mm diam and 32.4 mm length impacting a rigid wall with an initial axial velocity of 0.227 mm/ps (ref. 3). The bilinear elastic-plastic material model is used with isotropic hardening. A 90" segment of this axisymmetric cross section is modeled using 972 solid elements, 3552 D. E, and 80 time steps with a fixed step size of 1.0 x lo4 s. None of the subroutines were compiled with the KSRl optimization option. The CPU time for the serial version is 6060 s. Figure 3 (ref. 3) shows the stages of deformation at different time periods. The CPU times for each of the major steps and their percentages are shown in Table 1. The unusually high percentage (68.2%) of the element matrix generation step may be due to the elasto-plastic material and lack of the optimization option. 3. PREDICTION OF PARALLEL NIKE3D PERFORMANCE There are four major steps in the parallelization of NIKE3D: parallel matrix decomposition, parallel element matrix generation, parallel global stiffness matrix assembly, and parallel forwardbackward substitution. The prediction is based on actual computations of

6 ORNLDWG 95-5859 1.0 ms 0.0 ms 3.0ms 4.0 ms Fig. 2. The pipe whip test problem. 5.0 ms

Table 1. Central processing unit (CPU) time for each major step in the serial version Pipe whip problem, CPU time (%) Bar impacting problem, CPU time (%) (46.4) 1224 s (20.2) Element matrix generation 1065 s (45.1) 4131 s (68.2) Global stiffness matrix assembly Forwardbackward substitution Others I Total 65 s (2.8) 1 294 s (4-8) 6060 s (100.0)

8 ORNLDWG 95-5860 0.000 ms 0.016 ms 0.032 ms 0.048 ms 0.064 ms 0.080 ms Fig. 3. The bar impacting wall test problem.

9 parallel matrix decomposition on the KSRl multiprocessor system at ORNL and of parallel element matrix generation, parallel global stiffness matrix assembly, and parallel forward/backward substitution on the BBN TC2000 multiprocessor system at LLNL. 3.1 PARALLEL MATRIX DECOMPOSITION Parallel matrix decomposition of pipe whip and bar impacting rigid wall test problems was tested on the KSRl multiprocessor system.* This parallelization can reduce the CPU time of matrix decomposition from 1096 to 270 s for the pipe whip test problems and 1224 to 53 s for the bar impacting test problems. The higher speed-up performance for the bar impacting test problem is due to the fact that the serial version does not implement the compiler s optimization option, while the parallel version does. 3 2 PARALLEL ELEMENT MATRIX GENERATION Parallel element matrix generation performance was tested on the BBN TC2000 multiprocessor system at LLNL.2 The speed-up performance is over 90 for 104 processors under interleaved shared-memory environments. NIKE3D is a vectorized finite element code which is different from the tested code on the BBN TC2000 multiprocessor system. The element matrix generation is a highly parallelized step for a regular finite element code. It is believed that the vectorized version of element matrix generation should have the same highly parallel performance, as well as the regular version. However, due to the higher overhead, the predicted parallel element matrix generation performance for the NIKE3D should be around 30 to 40 for 64 processors on the KSRl system, depending on the implementation of the compiler s optimization option for the serial version. The predicted CPU time for the parallel element matrix generation should be around 36 s for the pipe whip test problem and 103 s for the bar impacting test problem. 3 3 PARALLEL GLOBAL STIFFNESS MATRIX ASSEMBLY The parallel global stiffness matrix assembly was tested on the BBN TC2000 multiprocessor system at LLNL.2 The speed-up performance is over 55 for 104 processors under interleaved shared-memory environments. The predicted parallel global stiffness matrix L assembly performance for the NIKE3D should be around 20 to 30 for 64 processors on the KSRl system, depending on the implementation of the compiler s optimization option for the

10 serial version. The predicted CPU time for the parallel global stiffness matrix assembly should be around 4 s for the pipe whip test problem and 10 s for the bar impacting test problem. 3.4 PARALLEL FORWARD/BACKWARDSUBSTITUTION Fornardbackward substitution with single right-hand side is not suitable for parallel computation due to the waiting steps in the forwardbackward substitution procedure.2 This forwardbackward step will keep in serial and have the same CPU time, that is, 69 s for the pipe whip test problem and 128 s for the bar impacting test problem. 3 5 PREDICI ION OFPARALLEL NIKE3D PERFORMANCE Tables 2 and 3 show the predicted parallel NIKE3D CPU time for each major step and overall performance with respect to the serial version for pipe whip and bar impacting test problems. There are some overheads when the parallel version of an algorithm is implemented on the serial version. This overhead may be up to around 50% of the serial version CPU time. The CPU time for parallel matrix decomposition with a single processor is 1659 s compared to 1096 s for the serial version in the pipe whip test problem. This situation will reduce the overall speed-up performance. The compiler s optimization option can reduce the tremendous CPU time required for serial computation on the KSRl system. It affects the CPU time and percentage for each step. The optimized serial version has a higher CPU time percentage for matrix decomposition, as the pipe whip test problem in Table 1 shows. The parallel version is considered to be optimized for the comparison. That is the reason why the bar impacting test problem has much higher speed-up performance than the pipe whip test problem. It is believed that the parallel performance of the bar impacting wall test problem will be in the same range as the pipe whip test problem if the compiler s optimization option is implemented in its serial version. The speed-up performance for banded matrix decomposition was limited. This is due to the fact that for locdshared-memory systems, there is a saturation point in the number of dedicited processors for any half bandwidth size.5 The performance remains the same above this saturation number of processors. The number of processors and the higher level of ring, ring:l, for the KSRl system may also affect the overall speed-up performance for NIKE3D. It is predicted that the overall parallel speed-up performance may be in the range of 5 to 10 for NIKE3D on the KSRl system with up to 64 processors. 4

11 Table 2. Predicted parallel performance for the pipe whip test problem on the KSRl system ~ Major step Parallel version, CPU time (%) Serial version, CPU time (%) Speed-up performance 270 s (60.8) 4.1 36 s (8.1) 29.6 4s (0.9) 16.8 Forwardbackward substitution 69 s (15.5) 1.0 Others 65 s (14.6) 1.o Matrix decomposition 1096 s (46.4) Element matrix generation 1065 s (45.1) 67 s (2.8) Global stiffness matrix assembly I 444 s (100.0) 2362 s (100.0) Total Table 3. Predicted parallel performance for the bar impacting wall test problem on the KSRl system Major step Matrix decomposition Element matrix generation Serial version, I Parallel version, CPU time (%) CPU time (%) 53 s (9.0) 23.1 103 s (17.5) 40.1 (1.7) 28.3 1224s (20-2) 4131 s (68.2) Global stiffness matrix assembly 283 s (4.7) Forwardbackward substitution 128s (2.1) 128 s (21.8) Others 294 s (4.8) 294 s (50.0) 6060 s (100.0) 588 s (100.0) Total I Speed-up performance 10 s I

12 4. CONCLUSIONS Predicted parallel performance of NIKE3D finite element code on the KSRl multiprocessor system has been presented in this paper. The results show that the parallel N I E 3 D is attractive for a multiprocessor computer architecture with local/shared-memory configuration. The compiler's optimization option can reduce the tremendous execution time of serial codes on the KSRl system, and it is highly recommended to implement this function. It is believed the results indicate that the parallel NIKE3D can provide an attractive strategy for high-performance finite element computations on multiprocessor, local/sharedmemory systems. The results can be used to predict the parallel performance of the commercial version of NIKE3D @.e.,ls-nike3d) on the local/shared-memory multiprocessor systems. Further experience with parallel sparse matrix storage algorithm would be desirable as well as the iterative method for equation solver. 5. REFERENCE3 1. P. S. Su,R. E. Fulton, and T. Zacharia, Implementation of Parallel Matrix Decomposition for NIKE3D on the KSRl System, ORNL/TM-12733, Martin Marietta Energy System, Inc., Oak Ridge Natl. Lab., December 1993. 2. P. S. Su, "Parallel Subdomain Method for Massively Parallel Computers," Ph.D. thesis, Georgia Institute of Technology, Atlanta, 1992. 3. B. N. Maker, R. M. Ferencz, and J. 0. Hallquist, NIKE3D:A Nonlinear, Implicit, Three-DimensionalFinite Element Code for Solid and Structural Mechanics - User's Manual, Lawrence Livermore Natl. Lab., Livermore, Calif., 1991. 4. Kendall Square Research, KSRl Principles of Operations, KSR 8/1/91, Rev. 5.5, Kenddll Square Research, Waltham, Mass., 1991. 5. R. E. Fulton, S. C. Chuang, P. S. Su, and S. Synn, "Structural Dynamics Methods for Parallel Supercomputers,"Phase V,Report for MacNeal-Schwendler Corporation, Georgia Institute of Technology, Atlanta, 1991.

13 ORNLlI M-12732 5 INTERNAL DISTRIBUTION 1 1-2. Central Research Library 3. Document Reference Section 4-5. Laboratory Records Department 6. Laboratory Records, ORNL RC 7. ORNL Patent Section 8-10. M&C Records Office 11. H. W. Hayden, Jr. 12. 13. 14. 15. 16. 17. E. Lara-Curzio S. Viswanathan H. W. Foglesong (Consultant) E. L. Menger (Consultant) J. G. Simon (Consultant) K, E. Spear (Consultant) EXTERNAL DISTRIBUTION 18. GEORGIA INSTITUTE OF TECHNOLOGY, School of Mechanical Engineering, Atlanta, GA 30332-0405 R.E. Fulton 19. KENDALL SQUARE RESEARCH CORPORATION, 170 Tracer Lane, Waltham, MA 02154-1379 M. Kaufman 20. LAWRENCE LIVERMORE NATIONAL LABORATORY, Method Development Group, P.O. Box 808, L-122, Livermore, CA 94550 B. N. Maker 21. LIVERMORE SOFTWARE TECHNOLOGY CORPORATION, 2876 Waverley Way, Livennore, CA 94550 J. 0. Hallquist 22. NASA LANGLEY, Research Center, Mail Stop 240, Hampton, VA 23681 0. 0. Storaasli 23. i UNIVERSITY OF TENNESSEE, JICS, Room 104, South College, Knoxville, TN 37996-1301 C. Halloy.

14 24-25 DOE, OAK RIDGE OPERATIONS OFFICE, P.O. Box 2001, Oak Ridge, TN 3783 1 E. E. Hoffman, National Materials Programs (MS 6269) Deputy Assistant Manager for Energy Research and Development (MS 6269) 26-27. DOE OFFICE OF SCIENTIFIC ANC TECHNICAL INFORMATION, P.O. Box 62, Oak Ridge, TN 37831 For distribution as shown in DOE/OSTI-4500, Distribution Category UC-404 (Materials)