Experiences of the Development of the Supercomputers
|
|
- Hubert Ira Berry
- 5 years ago
- Views:
Transcription
1 Experiences of the Development of the Supercomputers - Earth Simulator and K Computer YOKOKAWA, Mitsuo Kobe University/RIKEN AICS
2 Application Oriented Systems Developed in Japan No.1 systems in TOP500 list Titan Tianhe-2 FY FY
3 Talk Outline Earth Simulator (ES) Motivation and Target Performance of ES Achievements in Science & Engineering by ES K Computer Objectives and Design Targets Overview Storage System Summary 2
4 Motivation We had faced important and serious problems in global natural phenomena which affect human activities in Understanding and predicting the phenomena is essential. Climate change (Global warming) Weather Forecast (Heavy rain, droughts) Earthquake and Its Mechanism Air Pollution (acid rain, ozone hole) Simulation models and computer resources in 1997 are insufficient in resolution to solve these large-scale complicated phenomena. We started the development of the Earth Simulator in
5 Target Performance of ES Resolution in atmospheric circulation models Model 1997 Target Regional Model Global Model (AGCM) km km ~1km 5-10km Courtesy of JMA AGCM simulation with 10km mesh Horizontal meshes : 4000 x 2000 Vertical layers : several 10s Time step : around 1/10 At least 1000 times performance (about 5Tflop/s sustained). ( Cf. Sustained performance of AGCM simulation on typical supercomputer is 4-6Gflops in ) 4
6 Development of ES Design Concepts Distributed memory parallel system with vector processors as for compute nodes Full crossbar network More than 32TFLOPS peak performance & 8TB main memory capacity Development Schedule (5-year project) Late-project leader: Miyoshi-sensei Operation started in March 11,
7 Storage System of ES Hierarchical System Storage in ES User volume /home /data Work volume 228TB 43.4TB 184.5TB 460TB User volume Work volume MDPS Disks + Tape archive HDD 250TB Tape Archive 1.92PB STK Tape Drive 60GB x 32,000 cartridges MDPS (Mass Data Processing System) Total Capacity 2.86PB 6
8 100% Statistics of the ES Node Usage ( ) Computing Resource Distribution based on Job Size 80% 60% 40% 20% Development of MDPS IPCC 2002/ / / / / / /10 Running(S+L) Pre-run / post-run processing Idle Eco Stopped 10.2% 15.9% more than 256node 128~ 255nod e 64~127node 23.7% 32~ 63node 21.1% less than 15node 13.9% 16~31node 15.1% As of FY ,000 18,000 16,000 14,000 12,000 10,000 8,000 6,000 4,000 2,000 0 Number of jobs Courtesy of JAMSTEC 7
9 Contribution to the IPCC by ES Global warming projections by climate modeling team The Nobel Peace Prize 2007 Fourth Assessment Report by IPCC (Intergovernmental Panel on Climate Change) (2007) Major outcomes Highest resolution coupled model Very likely attribution for global warming (stronger confirmation) Super-high resolution global atmospheric model Projection of increased strength of typhoons and hurricanes (new finding) Earth system model Carbon cycle feedback causing additional warming (new finding) Working Group III (Mitigation) Working Group II (IAV*) Working Group I (Physical Science Basis) (* IAV :Impact, Adaptation and Vulnerability) Bali Roadmap (The Bali Climate Change Conference in 2007) 8
10 Seamless Simulation from Globe to City by MSSG Multi-scale Simulator for the Geo-environment (MSSG) A coupled model with a nonhydrostatic atmospheric general circulation model, a land model, a non-hydrostatic/hydrostatic ocean model, and an ocean wave model. Yin-Yang grids Grid Size 11 km Courtesy of Dr. Takahashi@JAMSTEC Grid size 5m 15 min. simulation required 4-5 hours on ES. 9
11 What did ES bring us? ES gave a big impact to HPC community with the word Computonik shock and a trigger to build more powerful supercomputers around the world. Simulation models in geoscience fields were extremely sophisticated. First-principle simulations in geoscience became more realistic. Possibility in predicting various phenomena by computer simulations was demonstrated. Necessity of more powerful supercomputers was shown. A development project of a next-generation supercomputer system K computer 10
12 Supercomputer K computer 京 [keɪ] is a distributed-memory parallel supercomputer developed by RIKEN in cooperation with Fujitsu Ltd. The nickname is after a target performance of the system, that is 10petaFLOPS (10 quadrillion floating point operations per second). 京 is a Japanese prefix number which stands for 10 16, or 10 peta. The K took first place twice in TOP500 supercomputer ranking list last year. Project Leader Dr. Tadashi Watanabe 11
13 Objectives and Design Targets (in 2006) Objectives to develop the world's most advanced and high-performance supercomputer, to develop and deploy its usage technologies including application software as one of Japan's Key Technologies of National Importance. Design Targets 10 peta-flops in LINPACK benchmark Peta-FLOPS sustained performance in various real applications Low power consumption system Highly reliable system The system should be a general-purpose machine and a production system for computational science. 12
14 Schedule of the Project and Its Location Start of installation at RIKEN AICS, Kobe End of installation No.1 in 38th TOP500 list (10.51 PFLOPS) HPCC Awards Gordon Bell Prize (Peak Performance) Completion of system development Open to academic and industry September 28, 2012 Operation started. 13
15 System Overview 14
16 Configuration of the K Computer Compute nodes (CPUs): 82,944 (IO nodes : 5,184) Peak performance: 10.6PFLOPS Memory: 1.27PB (16GB/node) 5GB/s(peak)x 2 6-dimensional mesh/torus network: Tofu 10 connections to adjacent compute nodes 24 * 18 * (16+1) * 2 * 3 * 2 Peak bandwidth: 5GB/s x 2 for each connection Logical 3-dimensional torus network SPARC64 TM VIIIfx 5GB/s(peak)x 2 ノード Compute node CPU: 128GFLOPS (8cores) Core Core SIMD(4FMA) Core SIMD(4FMA) Core SIMD(4FMA) Core SIMD(4FMA) Core SIMD(4FMA) Core16GFlops SIMD(4FMA) Core16GFlops 16GFlops SIMD(4FMA) 16GFlops SIMD(4FMA) 16GFlops 16GFlops 16GFLOPS 16GFlops L2$: 6MB 64GB/s MEM: 16GB 5GB/s(peak)x 2 5GB/s(peak)x 2 Courtesy of FUJITSU Ltd. 15
17 Fujitsu SPARC64 TM VIIIfx Chip Overview 8 cores 4 FMA (multiply and add) operation circuits in each core A SIMD instruction can execute 2 FMAs concurrently 16 giga-flops/core by 2 SIMDs x 2 FMAs in 2GHz 256 FP registers (double precision) Hardware barrier for fast inter-core synchronization Pre-fetch instruction for latency hiding Shared 6MB L2 Cache (12-way) Software controllable cache (sectored cache) Performance: 128 giga-flops/chip Reference: SPARC64 TM VIIIfx Extensions 45nm CMOS process, 2GHz 22.7mm x 22.6mm 760 M transistors 58W (at 30 by water cooling) 16
18 Tofu (Torus Fusion) Interconnect High communication performance and fault-tolerant network Six-dimensional mesh/torus network Each node has 10 links (5 GB/s x 2 bandwidth) 6 links for XYZ link (XZ:torus / Y:mesh) and 4 links for node link within a basic unit (2x3x2 mesh / torus) Multi-path routing by a combination of XYZ link and 2x3x2 mesh/torus network enables to make a detour of faulty nodes Z+ Y+ Basic unit (2x3x2 mesh / torus) Z Y X X- X+ Y- Z- Neighboring basic units are connected by 12 links. 17
19 System Environments Linux based operating system (OS) on compute nodes Batch job-oriented system Interactive environments are available for debugging. Two file systems of both local file system and global file system by FEFS (Fujitsu Exabyte File System) based on Lustre file system Users permanent files are on the global file system. Staging function Files on the global file system which are used in a job are staged into the local file system. Data generated during a job execution are moved back to the global file system after the job finished. 18
20 Batch Job Processing and Staging Function Job submission File staging Job execution File de-staging Users submit their jobs with directions to a scheduler. Input files used in the job are copied from a global file system to a local file system. The job is executed on the assigned compute nodes. Output files are moved back to the global file system Job # of nodes, expected elapsed time, staging, etc. in a job script Check the job status Job termination Check the job status Job submission Scheduler assigns the job to appropriate compute nodes according to the job Scheduler description Scheduler in the script. Scheduler Scheduler Compute nodes to which the job assigned Compute nodes Compute nodes Job execution Compute nodes Local File Global files ローカルファイル Input data Input data Global Files Input data ローカルファイル Output data Global Files ローカルファイル Output data Output data Global Files 19
21 From Chip to Full System 40m 40m Full System Compute Rack 864 Disk Rack cm 80cm 80cm 80cm 2 Cabinets Compute Rack 4 Disk Rack 1 Compute Rack SB 24 IOSB (11.3) PFLOPS 1.27 (1.34) PiB 50cm 50cm System Board(SB) Node 4 Node CPU ICC Memory 16GB 49.2 (52.4) TFLOPS 6.00 (6.38) TiB 128GFLOPS 16GiB 512GFLOPS 64GiB 12.3 (13.1) TFLOPS 1.50 (1.59) TiB Water cooling module Logical 3-D Torus 200,000 copper cables About 1,000km (620mi.) Courtesy of FUJITSU Ltd. 20
22 I/O Configuration Compute node racks Fibre Channel Local disk rack ~1.7GB/s InfiniBand 4X QDR Global File System ~720GB/s OSS nodes (90) ~3.2TB/s InfiniBand Switch 30PB~ 21
23 System Disks Compute node rack has storages for system files such as Linux kernel, log files, etc. 864 compute node racks in total RAID units / rack RAID5 (4D+1P) x 2sets + 2HS = in. 2TB SATA # of total disks 10,368 Compute node rack 22
24 Local File System (FEFS) Configuration MDS (Meta Data Server) 2 MDT (Meta Data Target) 4 RAID6 (6D+2P) x 4sets + 2HS = in. 2TB SATA 136 OSS (Object Storage Server) 648 OST (Object Storage Target) 2,592 RAID5 (4D+1P) x 4sets + 2HS = in. 300GB SAS 57,024 # of total disks 57,160 23
25 Global File System (FEFS) Configuration MDS (Meta Data Server) 2 MDT (Meta Data Target) 24 RAID1+0 (4D+4M) x 14sets + 7HS = in. 2TB SATA 2856 OSS (Object Storage Server) 90 OST (Object Storage Target) 720 RAID6(6D+2P) x 4sets + 2HS = in. 2TB SATA 24,480 # of total disks 27,336 24
26 Tsunami simulation of the Japan Earthquake on March 11, 2010 Computed Area:1192x768x500km 3 # of lattices:2304x1536x2000 Maeda et al., Bull. Seism. Soc. Am. in press,
27 Summary HPC and Big Data are essential for sustainable human life in the future and we have to promote HPC activities more and more. Storage technology is one of the keys to realize a large scale storage equipment Efforts for increasing reliability of those devices are critical for the future applications. 26
28 27
Current Status of the Next- Generation Supercomputer in Japan. YOKOKAWA, Mitsuo Next-Generation Supercomputer R&D Center RIKEN
Current Status of the Next- Generation Supercomputer in Japan YOKOKAWA, Mitsuo Next-Generation Supercomputer R&D Center RIKEN International Workshop on Peta-Scale Computing Programming Environment, Languages
More informationFujitsu s Approach to Application Centric Petascale Computing
Fujitsu s Approach to Application Centric Petascale Computing 2 nd Nov. 2010 Motoi Okuda Fujitsu Ltd. Agenda Japanese Next-Generation Supercomputer, K Computer Project Overview Design Targets System Overview
More informationIntroduction to the K computer
Introduction to the K computer Fumiyoshi Shoji Deputy Director Operations and Computer Technologies Div. Advanced Institute for Computational Science RIKEN Outline ü Overview of the K
More informationThe way toward peta-flops
The way toward peta-flops ISC-2011 Dr. Pierre Lagier Chief Technology Officer Fujitsu Systems Europe Where things started from DESIGN CONCEPTS 2 New challenges and requirements! Optimal sustained flops
More informationAn Overview of Fujitsu s Lustre Based File System
An Overview of Fujitsu s Lustre Based File System Shinji Sumimoto Fujitsu Limited Apr.12 2011 For Maximizing CPU Utilization by Minimizing File IO Overhead Outline Target System Overview Goals of Fujitsu
More informationFujitsu Petascale Supercomputer PRIMEHPC FX10. 4x2 racks (768 compute nodes) configuration. Copyright 2011 FUJITSU LIMITED
Fujitsu Petascale Supercomputer PRIMEHPC FX10 4x2 racks (768 compute nodes) configuration PRIMEHPC FX10 Highlights Scales up to 23.2 PFLOPS Improves Fujitsu s supercomputer technology employed in the FX1
More informationJapan HPC Programs - The Japanese national project of the K computer -
TERATEC 2011 Forum Japan HPC Programs - The Japanese national project of the K computer - June 28,2011 Tadashi WATANABE Next-Generation Supercomputer R&D Center RIKEN 0 Contents Outline of the Project
More informationThe Earth Simulator System
Architecture and Hardware for HPC Special Issue on High Performance Computing The Earth Simulator System - - - & - - - & - By Shinichi HABATA,* Mitsuo YOKOKAWA and Shigemune KITAWAKI The Earth Simulator,
More informationWhite paper Advanced Technologies of the Supercomputer PRIMEHPC FX10
White paper Advanced Technologies of the Supercomputer PRIMEHPC FX10 Next Generation Technical Computing Unit Fujitsu Limited Contents Overview of the PRIMEHPC FX10 Supercomputer 2 SPARC64 TM IXfx: Fujitsu-Developed
More informationThe Architecture and the Application Performance of the Earth Simulator
The Architecture and the Application Performance of the Earth Simulator Ken ichi Itakura (JAMSTEC) http://www.jamstec.go.jp 15 Dec., 2011 ICTS-TIFR Discussion Meeting-2011 1 Location of Earth Simulator
More informationThe Earth Simulator Current Status
The Earth Simulator Current Status SC13. 2013 Ken ichi Itakura (Earth Simulator Center, JAMSTEC) http://www.jamstec.go.jp 2013 SC13 NEC BOOTH PRESENTATION 1 JAMSTEC Organization Japan Agency for Marine-Earth
More informationPost-K Development and Introducing DLU. Copyright 2017 FUJITSU LIMITED
Post-K Development and Introducing DLU 0 Fujitsu s HPC Development Timeline K computer The K computer is still competitive in various fields; from advanced research to manufacturing. Deep Learning Unit
More informationWhite paper FUJITSU Supercomputer PRIMEHPC FX100 Evolution to the Next Generation
White paper FUJITSU Supercomputer PRIMEHPC FX100 Evolution to the Next Generation Next Generation Technical Computing Unit Fujitsu Limited Contents FUJITSU Supercomputer PRIMEHPC FX100 System Overview
More informationIntroduction of Fujitsu s next-generation supercomputer
Introduction of Fujitsu s next-generation supercomputer MATSUMOTO Takayuki July 16, 2014 HPC Platform Solutions Fujitsu has a long history of supercomputing over 30 years Technologies and experience of
More informationThe next generation supercomputer. Masami NARITA, Keiichi KATAYAMA Numerical Prediction Division, Japan Meteorological Agency
The next generation supercomputer and NWP system of JMA Masami NARITA, Keiichi KATAYAMA Numerical Prediction Division, Japan Meteorological Agency Contents JMA supercomputer systems Current system (Mar
More informationTechnical Computing Suite supporting the hybrid system
Technical Computing Suite supporting the hybrid system Supercomputer PRIMEHPC FX10 PRIMERGY x86 cluster Hybrid System Configuration Supercomputer PRIMEHPC FX10 PRIMERGY x86 cluster 6D mesh/torus Interconnect
More informationFujitsu s Technologies to the K Computer
Fujitsu s Technologies to the K Computer - a journey to practical Petascale computing platform - June 21 nd, 2011 Motoi Okuda FUJITSU Ltd. Agenda The Next generation supercomputer project of Japan The
More informationBrand-New Vector Supercomputer
Brand-New Vector Supercomputer NEC Corporation IT Platform Division Shintaro MOMOSE SC13 1 New Product NEC Released A Brand-New Vector Supercomputer, SX-ACE Just Now. Vector Supercomputer for Memory Bandwidth
More informationFujitsu s new supercomputer, delivering the next step in Exascale capability
Fujitsu s new supercomputer, delivering the next step in Exascale capability Toshiyuki Shimizu November 19th, 2014 0 Past, PRIMEHPC FX100, and roadmap for Exascale 2011 2012 2013 2014 2015 2016 2017 2018
More informationAdvanced Software for the Supercomputer PRIMEHPC FX10. Copyright 2011 FUJITSU LIMITED
Advanced Software for the Supercomputer PRIMEHPC FX10 System Configuration of PRIMEHPC FX10 nodes Login Compilation Job submission 6D mesh/torus Interconnect Local file system (Temporary area occupied
More informationFujitsu HPC Roadmap Beyond Petascale Computing. Toshiyuki Shimizu Fujitsu Limited
Fujitsu HPC Roadmap Beyond Petascale Computing Toshiyuki Shimizu Fujitsu Limited Outline Mission and HPC product portfolio K computer*, Fujitsu PRIMEHPC, and the future K computer and PRIMEHPC FX10 Post-FX10,
More informationJapan s post K Computer Yutaka Ishikawa Project Leader RIKEN AICS
Japan s post K Computer Yutaka Ishikawa Project Leader RIKEN AICS HPC User Forum, 7 th September, 2016 Outline of Talk Introduction of FLAGSHIP2020 project An Overview of post K system Concluding Remarks
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Waiting for Moore s Law to save your serial code start getting bleak in 2004 Source: published SPECInt data Moore s Law is not at all
More informationFujitsu s Technologies Leading to Practical Petascale Computing: K computer, PRIMEHPC FX10 and the Future
Fujitsu s Technologies Leading to Practical Petascale Computing: K computer, PRIMEHPC FX10 and the Future November 16 th, 2011 Motoi Okuda Technical Computing Solution Unit Fujitsu Limited Agenda Achievements
More informationFujitsu's Lustre Contributions - Policy and Roadmap-
Lustre Administrators and Developers Workshop 2014 Fujitsu's Lustre Contributions - Policy and Roadmap- Shinji Sumimoto, Kenichiro Sakai Fujitsu Limited, a member of OpenSFS Outline of This Talk Current
More informationOverview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo
Overview of Supercomputer Systems Supercomputing Division Information Technology Center The University of Tokyo Supercomputers at ITC, U. of Tokyo Oakleaf-fx (Fujitsu PRIMEHPC FX10) Total Peak performance
More informationChallenges in Developing Highly Reliable HPC systems
Dec. 1, 2012 JS International Symopsium on DVLSI Systems 2012 hallenges in Developing Highly Reliable HP systems Koichiro akayama Fujitsu Limited K computer Developed jointly by RIKEN and Fujitsu First
More informationPractical Scientific Computing
Practical Scientific Computing Performance-optimized Programming Preliminary discussion: July 11, 2008 Dr. Ralf-Peter Mundani, mundani@tum.de Dipl.-Ing. Ioan Lucian Muntean, muntean@in.tum.de MSc. Csaba
More informationThe Road from Peta to ExaFlop
The Road from Peta to ExaFlop Andreas Bechtolsheim June 23, 2009 HPC Driving the Computer Business Server Unit Mix (IDC 2008) Enterprise HPC Web 100 75 50 25 0 2003 2008 2013 HPC grew from 13% of units
More informationHOKUSAI System. Figure 0-1 System diagram
HOKUSAI System October 11, 2017 Information Systems Division, RIKEN 1.1 System Overview The HOKUSAI system consists of the following key components: - Massively Parallel Computer(GWMPC,BWMPC) - Application
More informationOverview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo
Overview of Supercomputer Systems Supercomputing Division Information Technology Center The University of Tokyo Supercomputers at ITC, U. of Tokyo Oakleaf-fx (Fujitsu PRIMEHPC FX10) Total Peak performance
More informationUser Training Cray XC40 IITM, Pune
User Training Cray XC40 IITM, Pune Sudhakar Yerneni, Raviteja K, Nachiket Manapragada, etc. 1 Cray XC40 Architecture & Packaging 3 Cray XC Series Building Blocks XC40 System Compute Blade 4 Compute Nodes
More informationThe Tofu Interconnect D
The Tofu Interconnect D 11 September 2018 Yuichiro Ajima, Takahiro Kawashima, Takayuki Okamoto, Naoyuki Shida, Kouichi Hirai, Toshiyuki Shimizu, Shinya Hiramoto, Yoshiro Ikeda, Takahide Yoshikawa, Kenji
More informationCluster Network Products
Cluster Network Products Cluster interconnects include, among others: Gigabit Ethernet Myrinet Quadrics InfiniBand 1 Interconnects in Top500 list 11/2009 2 Interconnects in Top500 list 11/2008 3 Cluster
More informationPost-K Supercomputer Overview. Copyright 2016 FUJITSU LIMITED
Post-K Supercomputer Overview 1 Post-K supercomputer overview Developing Post-K as the successor to the K computer with RIKEN Developing HPC-optimized high performance CPU and system software Selected
More informationGreen Supercomputing
Green Supercomputing On the Energy Consumption of Modern E-Science Prof. Dr. Thomas Ludwig German Climate Computing Centre Hamburg, Germany ludwig@dkrz.de Outline DKRZ 2013 and Climate Science The Exascale
More informationMIMD Overview. Intel Paragon XP/S Overview. XP/S Usage. XP/S Nodes and Interconnection. ! Distributed-memory MIMD multicomputer
MIMD Overview Intel Paragon XP/S Overview! MIMDs in the 1980s and 1990s! Distributed-memory multicomputers! Intel Paragon XP/S! Thinking Machines CM-5! IBM SP2! Distributed-memory multicomputers with hardware
More informationOverview of Supercomputer Systems. Supercomputing Division Information Technology Center The University of Tokyo
Overview of Supercomputer Systems Supercomputing Division Information Technology Center The University of Tokyo Supercomputers at ITC, U. of Tokyo Oakleaf-fx (Fujitsu PRIMEHPC FX10) Total Peak performance
More informationPRIMEHPC FX10: Advanced Software
PRIMEHPC FX10: Advanced Software Koh Hotta Fujitsu Limited System Software supports --- Stable/Robust & Low Overhead Execution of Large Scale Programs Operating System File System Program Development for
More informationGodson Processor and its Application in High Performance Computers
Godson Processor and its Application in High Performance Computers Weiwu Hu Institute of Computing Technology, Chinese Academy of Sciences Loongson Technologies Corporation Limited hww@ict.ac.cn 1 Contents
More informationOverview of Tianhe-2
Overview of Tianhe-2 (MilkyWay-2) Supercomputer Yutong Lu School of Computer Science, National University of Defense Technology; State Key Laboratory of High Performance Computing, China ytlu@nudt.edu.cn
More informationBlueGene/L. Computer Science, University of Warwick. Source: IBM
BlueGene/L Source: IBM 1 BlueGene/L networking BlueGene system employs various network types. Central is the torus interconnection network: 3D torus with wrap-around. Each node connects to six neighbours
More informationPost-Petascale Computing. Mitsuhisa Sato
Challenges on Programming Models and Languages for Post-Petascale Computing -- from Japanese NGS project "The K computer" to Exascale computing -- Mitsuhisa Sato Center for Computational Sciences (CCS),
More informationFindings from real petascale computer systems with meteorological applications
15 th ECMWF Workshop Findings from real petascale computer systems with meteorological applications Toshiyuki Shimizu Next Generation Technical Computing Unit FUJITSU LIMITED October 2nd, 2012 Outline
More informationFUJITSU HPC and the Development of the Post-K Supercomputer
FUJITSU HPC and the Development of the Post-K Supercomputer Toshiyuki Shimizu Vice President, System Development Division, Next Generation Technical Computing Unit 0 November 16 th, 2016 Post-K is currently
More informationDAQ system at SACLA and future plan for SPring-8-II
DAQ system at SACLA and future plan for SPring-8-II Takaki Hatsui T. Kameshima, Nakajima T. Abe, T. Sugimoto Y. Joti, M.Yamaga RIKEN SPring-8 Center IFDEPS 1 Evolution of Computing infrastructure from
More informationCommunication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.
Cluster Networks Introduction Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems. As usual, the driver is performance
More informationPerformance Evaluation of a Vector Supercomputer SX-Aurora TSUBASA
Performance Evaluation of a Vector Supercomputer SX-Aurora TSUBASA Kazuhiko Komatsu, S. Momose, Y. Isobe, O. Watanabe, A. Musa, M. Yokokawa, T. Aoyama, M. Sato, H. Kobayashi Tohoku University 14 November,
More informationCray XD1 Supercomputer Release 1.3 CRAY XD1 DATASHEET
CRAY XD1 DATASHEET Cray XD1 Supercomputer Release 1.3 Purpose-built for HPC delivers exceptional application performance Affordable power designed for a broad range of HPC workloads and budgets Linux,
More informationMathematical computations with GPUs
Master Educational Program Information technology in applications Mathematical computations with GPUs Introduction Alexey A. Romanenko arom@ccfit.nsu.ru Novosibirsk State University How to.. Process terabytes
More informationMAHA. - Supercomputing System for Bioinformatics
MAHA - Supercomputing System for Bioinformatics - 2013.01.29 Outline 1. MAHA HW 2. MAHA SW 3. MAHA Storage System 2 ETRI HPC R&D Area - Overview Research area Computing HW MAHA System HW - Rpeak : 0.3
More informationThe Cray Rainier System: Integrated Scalar/Vector Computing
THE SUPERCOMPUTER COMPANY The Cray Rainier System: Integrated Scalar/Vector Computing Per Nyberg 11 th ECMWF Workshop on HPC in Meteorology Topics Current Product Overview Cray Technology Strengths Rainier
More informationFujitsu and the HPC Pyramid
Fujitsu and the HPC Pyramid Wolfgang Gentzsch Executive HPC Strategist (external) Fujitsu Global HPC Competence Center June 20 th, 2012 1 Copyright 2012 FUJITSU "Fujitsu's objective is to contribute to
More informationSystem Software for Big Data and Post Petascale Computing
The Japanese Extreme Big Data Workshop February 26, 2014 System Software for Big Data and Post Petascale Computing Osamu Tatebe University of Tsukuba I/O performance requirement for exascale applications
More informationFujitsu High Performance CPU for the Post-K Computer
Fujitsu High Performance CPU for the Post-K Computer August 21 st, 2018 Toshio Yoshida FUJITSU LIMITED 0 Key Message A64FX is the new Fujitsu-designed Arm processor It is used in the post-k computer A64FX
More informationPerformance Evaluation of TOFU System Area Network Design for High- Performance Computer Systems
Performance Evaluation of TOFU System Area Network Design for High- Performance Computer Systems P. BOROVSKA, O. NAKOV, S. MARKOV, D. IVANOVA, F. FILIPOV Computer System Department Technical University
More informationPost-K: Building the Arm HPC Ecosystem
Post-K: Building the Arm HPC Ecosystem Toshiyuki Shimizu FUJITSU LIMITED Nov. 14th, 2017 Exhibitor Forum, SC17, Nov. 14, 2017 0 Post-K: Building up Arm HPC Ecosystem Fujitsu s approach for HPC Approach
More informationAim High. Intel Technical Update Teratec 07 Symposium. June 20, Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group
Aim High Intel Technical Update Teratec 07 Symposium June 20, 2007 Stephen R. Wheat, Ph.D. Director, HPC Digital Enterprise Group Risk Factors Today s s presentations contain forward-looking statements.
More informationSun Lustre Storage System Simplifying and Accelerating Lustre Deployments
Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments Torben Kling-Petersen, PhD Presenter s Name Principle Field Title andengineer Division HPC &Cloud LoB SunComputing Microsystems
More informationFabio AFFINITO.
Introduction to High Performance Computing Fabio AFFINITO What is the meaning of High Performance Computing? What does HIGH PERFORMANCE mean??? 1976... Cray-1 supercomputer First commercial successful
More informationAccelerating High Performance Computing.
Accelerating High Performance Computing http://www.nvidia.com/tesla Computing The 3 rd Pillar of Science Drug Design Molecular Dynamics Seismic Imaging Reverse Time Migration Automotive Design Computational
More informationTofu Interconnect 2: System-on-Chip Integration of High-Performance Interconnect
Tofu Interconnect 2: System-on-Chip Integration of High-Performance Interconnect Yuichiro Ajima, Tomohiro Inoue, Shinya Hiramoto, Shunji Uno, Shinji Sumimoto, Kenichi Miura, Naoyuki Shida, Takahiro Kawashima,
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist Waiting for Moore s Law to save your serial code started getting bleak in 2004 Source: published SPECInt
More informationPractical Scientific Computing
Practical Scientific Computing Performance-optimised Programming Preliminary discussion, 17.7.2007 Dr. Ralf-Peter Mundani, mundani@tum.de Dipl.-Ing. Ioan Lucian Muntean, muntean@in.tum.de Dipl.-Geophys.
More informationGoro Watanabe. Bill King. OOW 2013 The Best Platform for Big Data and Oracle Database 12c. EVP Fujitsu R&D Center North America
OOW 2013 The Best Platform for Big Data and Oracle Database 12c Goro Watanabe EVP Fujitsu R&D Center North America Bill King EVP Platform Products Group Fujitsu America, Inc. Overview 1. Fujitsu: Quick
More informationComplexity and Advanced Algorithms. Introduction to Parallel Algorithms
Complexity and Advanced Algorithms Introduction to Parallel Algorithms Why Parallel Computing? Save time, resources, memory,... Who is using it? Academia Industry Government Individuals? Two practical
More informationHPC and Big Data: Updates about China. Haohuan FU August 29 th, 2017
HPC and Big Data: Updates about China Haohuan FU August 29 th, 2017 1 Outline HPC and Big Data Projects in China Recent Efforts on Tianhe-2 Recent Efforts on Sunway TaihuLight 2 MOST HPC Projects 2016
More informationIntroduction to Parallel and Distributed Computing. Linh B. Ngo CPSC 3620
Introduction to Parallel and Distributed Computing Linh B. Ngo CPSC 3620 Overview: What is Parallel Computing To be run using multiple processors A problem is broken into discrete parts that can be solved
More informationArchitecting Storage for Semiconductor Design: Manufacturing Preparation
White Paper Architecting Storage for Semiconductor Design: Manufacturing Preparation March 2012 WP-7157 EXECUTIVE SUMMARY The manufacturing preparation phase of semiconductor design especially mask data
More informationHybrid Architectures Why Should I Bother?
Hybrid Architectures Why Should I Bother? CSCS-FoMICS-USI Summer School on Computer Simulations in Science and Engineering Michael Bader July 8 19, 2013 Computer Simulations in Science and Engineering,
More informationLustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE
Lustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE Hitoshi Sato *1, Shuichi Ihara *2, Satoshi Matsuoka *1 *1 Tokyo Institute
More informationHigh Performance Computing
CSC630/CSC730: Parallel & Distributed Computing Trends in HPC 1 High Performance Computing High-performance computing (HPC) is the use of supercomputers and parallel processing techniques for solving complex
More informationin Action Fujitsu High Performance Computing Ecosystem Human Centric Innovation Innovation Flexibility Simplicity
Fujitsu High Performance Computing Ecosystem Human Centric Innovation in Action Dr. Pierre Lagier Chief Technology Officer Fujitsu Systems Europe Innovation Flexibility Simplicity INTERNAL USE ONLY 0 Copyright
More informationLeibniz Supercomputer Centre. Movie on YouTube
SuperMUC @ Leibniz Supercomputer Centre Movie on YouTube Peak Performance Peak performance: 3 Peta Flops 3*10 15 Flops Mega 10 6 million Giga 10 9 billion Tera 10 12 trillion Peta 10 15 quadrillion Exa
More informationShared Services Canada Environment and Climate Change Canada HPC Renewal Project
Shared Services Canada Environment and Climate Change Canada HPC Renewal Project CUG 2017 Redmond, WA, USA Deric Sullivan Alain St-Denis & Luc Corbeil May 2017 Background: SSC's HPC Renewal for ECCC Environment
More informationFeedback on BeeGFS. A Parallel File System for High Performance Computing
Feedback on BeeGFS A Parallel File System for High Performance Computing Philippe Dos Santos et Georges Raseev FR 2764 Fédération de Recherche LUmière MATière December 13 2016 LOGO CNRS LOGO IO December
More informationA Multicore Processor Designed For PetaFLOPS Computation
A Multicore Processor Designed For PetaFLOPS Computation Weiwu Hu Institute of Computing Technology, Chinese Academy of Sciences Loongson Technologies Corporation Limited hww@ict.ac.cn 1 Contents Background
More informationImplicit Low-Order Unstructured Finite-Element Multiple Simulation Enhanced by Dense Computation using OpenACC
Fourth Workshop on Accelerator Programming Using Directives (WACCPD), Nov. 13, 2017 Implicit Low-Order Unstructured Finite-Element Multiple Simulation Enhanced by Dense Computation using OpenACC Takuma
More informationReal Parallel Computers
Real Parallel Computers Modular data centers Overview Short history of parallel machines Cluster computing Blue Gene supercomputer Performance development, top-500 DAS: Distributed supercomputing Short
More informationHigh Performance Computing. What is it used for and why?
High Performance Computing What is it used for and why? Overview What is it used for? Drivers for HPC Examples of usage Why do you need to learn the basics? Hardware layout and structure matters Serial
More informationAtos announces the Bull sequana X1000 the first exascale-class supercomputer. Jakub Venc
Atos announces the Bull sequana X1000 the first exascale-class supercomputer Jakub Venc The world is changing The world is changing Digital simulation will be the key contributor to overcome 21 st century
More informationTitan - Early Experience with the Titan System at Oak Ridge National Laboratory
Office of Science Titan - Early Experience with the Titan System at Oak Ridge National Laboratory Buddy Bland Project Director Oak Ridge Leadership Computing Facility November 13, 2012 ORNL s Titan Hybrid
More informationBlue Gene/Q. Hardware Overview Michael Stephan. Mitglied der Helmholtz-Gemeinschaft
Blue Gene/Q Hardware Overview 02.02.2015 Michael Stephan Blue Gene/Q: Design goals System-on-Chip (SoC) design Processor comprises both processing cores and network Optimal performance / watt ratio Small
More informationHigh-Performance Lustre with Maximum Data Assurance
High-Performance Lustre with Maximum Data Assurance Silicon Graphics International Corp. 900 North McCarthy Blvd. Milpitas, CA 95035 Disclaimer and Copyright Notice The information presented here is meant
More informationAn Empirical Study of Computation-Intensive Loops for Identifying and Classifying Loop Kernels
An Empirical Study of Computation-Intensive Loops for Identifying and Classifying Loop Kernels Masatomo Hashimoto Masaaki Terai Toshiyuki Maeda Kazuo Minami 26/04/2017 ICPE2017 1 Agenda Performance engineering
More informationSupercomputing at the United States National Weather Service (NWS)
Supercomputing at the United States National Weather Service (NWS) Rebecca Cosgrove Deputy Director, NCEP Central Operations United States National Weather Service 18th Workshop on HPC in Meteorology September
More informationResources Current and Future Systems. Timothy H. Kaiser, Ph.D.
Resources Current and Future Systems Timothy H. Kaiser, Ph.D. tkaiser@mines.edu 1 Most likely talk to be out of date History of Top 500 Issues with building bigger machines Current and near future academic
More informationJust on time to face new challenges with NEC super-computer at Meteo-France
Just on time to face new challenges with NEC super-computer at Meteo-France Agenda of the procurement Presentation of the first phase Installation phase (power supply, air cooling) Use of a porting machine
More informationThe current status of the adoption of ZFS* as backend file system for Lustre*: an early evaluation
The current status of the adoption of ZFS as backend file system for Lustre: an early evaluation Gabriele Paciucci EMEA Solution Architect Outline The goal of this presentation is to update the current
More informationTrends in HPC (hardware complexity and software challenges)
Trends in HPC (hardware complexity and software challenges) Mike Giles Oxford e-research Centre Mathematical Institute MIT seminar March 13th, 2013 Mike Giles (Oxford) HPC Trends March 13th, 2013 1 / 18
More informationEN2910A: Advanced Computer Architecture Topic 06: Supercomputers & Data Centers Prof. Sherief Reda School of Engineering Brown University
EN2910A: Advanced Computer Architecture Topic 06: Supercomputers & Data Centers Prof. Sherief Reda School of Engineering Brown University Material from: The Datacenter as a Computer: An Introduction to
More informationGetting the best performance from massively parallel computer
Getting the best performance from massively parallel computer June 6 th, 2013 Takashi Aoki Next Generation Technical Computing Unit Fujitsu Limited Agenda Second generation petascale supercomputer PRIMEHPC
More informationHigh-density Grid storage system optimization at ASGC. Shu-Ting Liao ASGC Operation team ISGC 2011
High-density Grid storage system optimization at ASGC Shu-Ting Liao ASGC Operation team ISGC 211 Outline Introduction to ASGC Grid storage system Storage status and issues in 21 Storage optimization Summary
More informationChallenges in High Performance Computing. William Gropp
Challenges in High Performance Computing William Gropp www.cs.illinois.edu/~wgropp 2 What is HPC? High Performance Computing is the use of computing to solve challenging problems that require significant
More informationIt s a Multicore World. John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist
It s a Multicore World John Urbanic Pittsburgh Supercomputing Center Parallel Computing Scientist Waiting for Moore s Law to save your serial code started getting bleak in 2004 Source: published SPECInt
More informationOutline. Execution Environments for Parallel Applications. Supercomputers. Supercomputers
Outline Execution Environments for Parallel Applications Master CANS 2007/2008 Departament d Arquitectura de Computadors Universitat Politècnica de Catalunya Supercomputers OS abstractions Extended OS
More informationCRAY XK6 REDEFINING SUPERCOMPUTING. - Sanjana Rakhecha - Nishad Nerurkar
CRAY XK6 REDEFINING SUPERCOMPUTING - Sanjana Rakhecha - Nishad Nerurkar CONTENTS Introduction History Specifications Cray XK6 Architecture Performance Industry acceptance and applications Summary INTRODUCTION
More informationThe Red Storm System: Architecture, System Update and Performance Analysis
The Red Storm System: Architecture, System Update and Performance Analysis Douglas Doerfler, Jim Tomkins Sandia National Laboratories Center for Computation, Computers, Information and Mathematics LACSI
More information1. ALMA Pipeline Cluster specification. 2. Compute processing node specification: $26K
1. ALMA Pipeline Cluster specification The following document describes the recommended hardware for the Chilean based cluster for the ALMA pipeline and local post processing to support early science and
More informationParallel Computer Architecture II
Parallel Computer Architecture II Stefan Lang Interdisciplinary Center for Scientific Computing (IWR) University of Heidelberg INF 368, Room 532 D-692 Heidelberg phone: 622/54-8264 email: Stefan.Lang@iwr.uni-heidelberg.de
More information