Presentation is loading. Please wait.

Presentation is loading. Please wait.

Updates of AICS and the next step for Post-Petascale Computing in Japan Mitsuhisa Sato University of Tsukuba Team leader of programming environment research.

Similar presentations


Presentation on theme: "Updates of AICS and the next step for Post-Petascale Computing in Japan Mitsuhisa Sato University of Tsukuba Team leader of programming environment research."— Presentation transcript:

1 Updates of AICS and the next step for Post-Petascale Computing in Japan Mitsuhisa Sato University of Tsukuba Team leader of programming environment research team, Advanced Institute for Computational Science (AICS), RIKEN 0

2 1 RIKEN Advanced Institute for Computational Science (AICS) The institute have been established at the K computer in Kobe (started in October 2010) Missions To run the K computer efficiently for users of wide research areas Carry out the leading edge of computational science technologies and contribute for COE of computational science in Japan Propose the future directions of HPC in Japan and conduct it. Organization Operation division, to run and manage the K computer Research division Started with 5 computational science research teams and 3 computer science research teams. In 2012, expanded to 10 computational science research teams and 6 computer science research teams. Promoting strong collaborations between computational and computer scientists, working with core-organizations of each fields together. http://www.aics.riken.jp/

3 Divisions AICS Policy Planning Division AICS Research Support Division Research Division Operations and Computer Technologies Division 2

4 Research Division (16teams + 3units) System Software Research Team Programming Environment Research Team Processor Research Team Large-scale Parallel Numerical Computing Technology Research Team HPC Usability Research Team Field Theory Research Team Discrete Event Simulation Research Team Computational Molecular Science Research Team Computational Materials Science Research Team Computational Biophysics Research Team Particle Simulator Research Team Computational Climate Science Research Team Complex Phenomena Unified Simulation Research Team HPC Programming Framework Research Team Advanced Visualization Research Team Data Assimilation Research Team Computational Chemistry Research Unit Computational Disaster Mitigation and Reduction Research Unit Computational Structural Biology Research Unit 3

5 The status of the K computer The first racks of the K computer were delivered to Kobe on Sept, 2010. Rack : 864 (+54), Compute nodes (CPUs): 82,944 (88,128), Number of cores: 663,552(705,024) It has already achieved its primary target “ over 10 Peta-flops” (10.51PF Linpack, 12.66MW), the last November (2011). Installation and adjustment of K was complete, and the public use was started at the end of the September (2012). Photo of First delivery, Sep 28, 2010

6 5 K computer: compute nodes and network Rack : 864 (+ 54 ) Compute nodes (CPUs): 82,944 (88128 ) Number of cores: 663,552 (705024 ) Peak performance: 10.6PFLOPS (11.28) 10.51PF Linpack (12.66MW), Nov 2011 Memory: 1.33 PB (16GB/node) SPARC64 TM VIIIfx Courtesy of FUJITSU Ltd. Compute node Logical 3-dimensional torus network for programming High-Performance/Low Power CPU with 8 cores : 128GFlops@2GHz, 58W High Throughput/Low Latency Torus Network (Tofu) Logical 3-dimensional torus network Peak bandwidth: 5GB/s x 2 for each direction of logical 3-dimensional torus network bi-section bandwidth: 30TB/s

7 6 京コンピュータ “The K computer"

8 7 Projects to organize users of the K computer SPIRE (Strategic Programs for Innovative Research) The committee in MEXT has identified five application areas that are expected to create breakthroughs using the K computer from national viewpoint. National-wide High Performance Computing Infrastructure Project (HPCI) To organize computing resources and users, including university supercomputers the K computing in national-wide

9 SPIRE (Strategic Programs for Innovative Research) Purpose To produce scientific results as soon as HPCI starts its operation To establish several core institutes for computational science Outline of this program Identify the five strategic research areas which will contribute to produce results to scientific and social Issues A nation wide research groups are formed by funding the core organization designated by MEXT. The groups are to promote R&D using K computer and to construct research structures for their own area 50% computing resources of the K computer will be dedicated to this program

10 Five strategic areas of SPIRE Life science/Drug manufacture Monodukuri (Manufacturing technology) Monodukuri (Manufacturing technology) New material/energy creation Global change prediction for disaster prevention/mitigation ゲノム 全身 タンパク質 細胞 多階層の生命現象 組織,臓器 Toshio YANAGIDA (RIKEN) Shinji TSUNEYUKI (University of Tokyo) Shiro IMAWAKI (JAMSTEC) Chisachi KATO (University of Tokyo) The origin of matter and the universe Shinya AOKI (University of Tsukuba)

11 National-wide High Performance Computing Infrastructure Project (HPCI) Background: After re-evaluation of the project at “government party change” in 2011, the NGS project was restarted as “Creation of the Innovative High Performance Computing Infra-structure (HPCI)”. Building HPCI: High-Performance Computing Infrastructure Provide Seamless access to K computer, supercomputers, and user's machines Set up a large-scale storage system for the K computer and other supercomputers Joint selection of proposals for K and other supercomputers Organizing HPCI Consortium Organize users and computer centers and provide proposals/suggestions to the government and related organizations Plan and operation of HPCI system Promotion of computational sciences Future supercomputing consortium super computer super computer super computer super computer HPCI K computer Institutional/University computer centers Computational Science communities Advanced Institute for Computational Science (AICS), RIKEN The consortium has been organized and started in June 2012 10

12 The Conceptual View of HPCI Consortium users K computer Large Scale Storage Supercomputers in Universities HPCI is a comprehensive advanced computing infrastructure in which the supercomputers and large scale storages are connected together through the high speed network.

13 Computing resources in HPCI 12

14 Storage System in first phase for HPCI Hokkaido University Tohoku University University of Tokyo University of Tsukuba Tokyo Institute of Technology Nagoya University Kyushu University Osaka UniversityKyoto University AICS, RIKEN 12 PB storage (52 OSS) Cluster for data analysis (87 node) 10 PB storage (30 OSS) Cluster for data analysis (88 node) HPCI WEST HUB HPCI EAST HUB Gfarm2 is used as the global shared file system 13

15 HPCI offers to computational science users … Computing resource Computational science researchers can make use of proper amount of computing resources more effectively and efficiently K computer and university supercomputers Authentication Users can get access to these computers and storages with a single- sign-on account Storage Users can share a large amount of data in the storage Analyze or visualize results simulated by other researchers on different supercomputers 14

16 Users and jobs about 100 active users and 1000 jobs per day period I Sep.2012-Mar.2013 period II Apr.2013-Sep.2013 period III Oct.2013-Mar.2014

17 Job Property(1/2) Larger jobs (more than 5000 nodes (= 0.5PF)) consume about 40-50% of the resource. Used / Serviced ratio reaches 80%. period I Sep.2012-Mar.2013 period II Apr.2013-Sep.2013 period III Oct.2013-Mar.2014

18 Job Property(2/2) Average of the sustained performances and job scale are gradually increasing.

19 Activities for the future HPC development FY2011FY2012 FY2013 SDHPC WGs: Technical discussion by two groups FS “Feasibility study” for future HPC Discussion on HPC policy Decision of future HPC R&D Discussion on basic concept SDHPC White paper was published Review We are here now

20 The SDHPC white paper and “Feasibility Study" project WGs were organized for drafting the white paper for Strategic Direction/Development of HPC in JAPAN by young Japanese researchers with advisers (seniors) Contents Science roadmap until 2020 and List of application for 2020’s Four types of hardware architectures identified and performance projection in 2018 estimated from the present technology trend Necessity of further research and development to realize the science roadmap For “Feasibility Study" project, 4 research teams were accepted Application study team leaded by RIKEN AICS (Tomita) System study team leaded by U Tokyo (Ishikawa) Next-generation “General-Purpose” Supercomputer System study team leaded by U Tsukuba (Sato) Study on exascale heterogeneous systems with accelerators System study team leaded by Tohoku U (Kobayashi) Projects were started from July 2012 (1.5 year) … 19

21 System requirement analysis for Target sciences System performance FLOPS: 800 – 2500PFLOPS Memory capacity: 10TB – 500PB Memory bandwidth: 0.001 – 1.0 B/F Example applications Small capacity requirement MD, Climate, Space physics, … Small BW requirement Quantum chemistry, … High capacity/BW requirement Incompressibility fluid dynamics, … Interconnection Network Not enough analysis has been carried out Some applications need >1us latency and large bisection BW Storage There is not so big demand 20 Low BW Middle capacity High BW small capacity High BW middle capacity High BW High capacity (From SDHPC white paper)

22 Alternatives of ExaScale Architecture Four types of architectures are identified for exascale: General Purpose (GP) Ordinary CPU-based MPPs e.g.) K-Computer, GPU, Blue Gene, x86-based PC-clusters Capacity-Bandwidth oriented (CB) With expensive memory-I/F rather than computing capability e.g.) Vector machines Reduced Memory (RM) With embedded (main) memory e.g.) SoC, MD-GRAPE4, Anton Compute Oriented (CO) Many processing units e.g.) ClearSpeed, GRAPE-DR, GPU? 21 (From SDHPC white paper)

23 22 Issues for exascale computing Two important aspects of post- petascale computing Power limitation < 20-30 MW Strong-scaling < 10^6 nodes, for FT > 10TFlops/node accelerator, many-cores Solution: Accelerated Computing by GPGPU by Application-specific Accelerator by... future acceleration device... the K computer simple projection of #nodes and peak flops

24 Study on exascale heterogeneous systems with accelerators (U Tsukuba project) Two keys for exascale computing Power and strong-scaling We study “exascale” heterogeneous systems with accelerators of many- cores. We are interested in: Architecture of accelerators, core and memory architecture Special-purpose functions Direct connection between accelerators in a group Power estimation and evaluation Programming model and computational science applications Requirement for general-purpose system etc … 23

25 PACS-G: a straw man architecture SIMD architecture, for compute oriented apps (N-body, MD), and stencil apps. 4096 cores (64x64), 2FMA@1GHz, 4GFlops x 4096 = 16TFlops/chip 2D mesh (+ broardcast/reduction) on-chip network for stencil apps. We expect 14nm technology at the range of year 2018-2020, Chip dai size: 20mm x 20mm Mainly working on on-chip memory (size 512 MB/chip, 128KB/core), and, with module memory by 3D-stack/wide IO DRAM memory (via 2.5D TSV), bandwidth 1000GB/s, size 16-32GB/chip No external memory (DIM/DDR) 250 W/chips expected 64K chips for 1 EFLOPS (at peak) PACS-G チップ 3D stack or Wide IO DRAM

26 A group of 1024 ~ 2048 chips are connected via accelerator network (inter-chip network) 25 – 50Gpbs/link for inter-chip: If we extend 2- D mesh network to the (2D-mesh) external net work in a group, we need 200 ~ 400GB/s (= 32 ch. x 25 ~ 50Gbps x 2(bi-direction)) For 50Gpbs data transfer, we may need direct optical interconnect from chip. I/O Interface to Host: PCI Express Gen 4 x16 (not enough!!!) interconnect between chips (2D mesh) Programming model: XcalableMP + OpenACC Use OpenACC to specify offloaded fragme nt of code and data movement To align data and computation to core, we use the concept "template" of XcalableMP (virtual index space). We can generate code for each core. (And data parallel lang. like C*) An example of implementation (for 1U rack) PACS-G: a straw man architecture

27 Project organization Joint project with Titech (Makino), Aizu U (Nakazato), RIKEN (Taiji), U Tokyo, KEK, Hiroshima U, and Hitachi as a super computer company Target apps: QCD in particle physics, tree N-body, HMD in Astrophysics, MD in life sci., FDM of earthquake, FMO in chemistry, NICAM in climate sci. 26

28 Current status and schedule We are now working on performance estimation by co-design process 2012 (done): QCD, N-body, MD, HMD 2013: earth quake sim, NICAM (climate), FMO (chemistry) When all data fits on on-chip memory, ratio B/F is 4 B/F, total mem size 1TB/group When data fits into module memory, ratio B/F is 0.05B/F, total mem size 32TB/group Also, developing simulators (clock-level/instruction level) for more precious and quantitative performance evaluation Compiler development (XMP and OpenACC) (Re-)Design and investigation of network topology 2D mesh is sufficient? or, other alternatives? Code development for apps using Host and Acc, including I/O Precise and more detail estimation of power consumptions

29 AICS Development of International Partnership NCSA, US under MoU (1 st meeting in April, 2013) NCI (National Computational Infracture), Australia, under MoU JSC, Germany, under MoU SISSA (THE SCUOLA INTERNAZIONALE SUPERIORE DI STUDI AVANZATI), Italia, under agreement. Unversity of Maryland, US, with agreement for collaboration on modeling and data assimilation. "maison de la simulation" (INRIA/CNRS), France, under discussion ANL, US, under discussion Recently, Japan and US agreed the collaboration on system software for supercomputing at "U.S.-Japan Joint High-Level Committee Meeting on Science and Technology Cooperation" http://www.state.gov/r/pa/prs/ps/2013/04/208651.htm Workshop on international collaboration for exascale computing will be organized in ISC2013, next week. Towards to JLESC. (esp. for exascale software development) 28


Download ppt "Updates of AICS and the next step for Post-Petascale Computing in Japan Mitsuhisa Sato University of Tsukuba Team leader of programming environment research."

Similar presentations


Ads by Google