GravitySimulator Beyond the Million Body Problem Collaborators:Rainer Spurzem (Heidelberg) Peter Berczik (Heidelberg/Kiev) Simon Portegies Zwart (Amsterdam)

Slides:



Advertisements
Similar presentations
Clusters, Grids and their applications in Physics David Barnes (Astro) Lyle Winton (EPP)
Advertisements

Spiegel - A Visualization System for Stellar Systems Hans-Peter David
Ido Tov & Matan Raveh Parallel Processing ( ) January 2014 Electrical and Computer Engineering DPT. Ben-Gurion University.
2. Computer Clusters for Scalable Parallel Computing
Contact: Hirofumi Amano at Kyushu 40 Years of HPC Services In this memorable year, the.
1 DISTRIBUTION STATEMENT XXX– Unclassified, Unlimited Distribution Laser Propagation Modeling Making Large Scale Ultrashort Pulse Laser Simulations Possible.
1 Coven a Framework for High Performance Problem Solving Environments Nathan A. DeBardeleben Walter B. Ligon III Sourabh Pandit Dan C. Stanzione Jr. Parallel.
Ryuji Morishima (UCLA/JPL). N-body code: Gravity solver + Integrator Gravity solver must be fast and handle close encounters Special hardware (N 2 ):
Petaflops Special-Purpose Computer for Molecular Dynamics Simulations Makoto Taiji High-Performance Molecular Simulation Team Computational & Experimental.
10/21/20091 Protein Explorer: A Petaflops Special-Purpose Computer System for Molecular Dynamics Simulations Makoto Taiji, Tetsu Narumi, Yousuke Ohno,
Galactic Super Massive Binary Black Hole Mergers Galactic Super Massive Binary Black Hole Mergers Dr. Peter Berczik Astronomisches Rechen-Institut (ARI),
Module on Computational Astrophysics Jim Stone Department of Astrophysical Sciences 125 Peyton Hall : ph :
1 June 2015 AstroGrid-D WP 5: Resource Management for Grid Jobs Report by: Rainer Spurzem (ZAH-ARI) and T. Brüsemeister,
Module on Computational Astrophysics Professor Jim Stone Department of Astrophysical Sciences and PACM.
Fabio Antonini Joshua Faber Alessia Gualandris and David Merritt Rochester Institute of Technology.
Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
Simon Portegies Zwart (Univ. Amsterdam with 2 GRAPE-6 boards)
PARALLEL PROCESSING The NAS Parallel Benchmarks Daniel Gross Chen Haiout.
Parallel Rendering Ed Angel
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.
Chapter 2 Computer Clusters Lecture 2.1 Overview.
CS 221 – May 13 Review chapter 1 Lab – Show me your C programs – Black spaghetti – connect remaining machines – Be able to ping, ssh, and transfer files.
1 Application of multiprocessor and GRID technology in medical image processing IKTA /2002.
Parallel Rendering 1. 2 Introduction In many situations, standard rendering pipeline not sufficient ­Need higher resolution display ­More primitives than.
Course Outline Introduction in algorithms and applications Parallel machines and architectures Overview of parallel machines, trends in top-500, clusters,
Low-Power Wireless Sensor Networks
Maximizing The Compute Power With Mellanox InfiniBand Connectivity Gilad Shainer Wolfram Technology Conference 2006.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Chapter 2 Computer Clusters Lecture 2.2 Computer Cluster Architectures.
Pursuing Faster I/O in COSMO POMPA Workshop May 3rd 2010.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Edgar Gabriel Short Course: Advanced programming with MPI Edgar Gabriel Spring 2007.
Parallelization: Area Under a Curve. AUC: An important task in science Neuroscience – Endocrine levels in the body over time Economics – Discounting:
2005 Materials Computation Center External Board Meeting The Materials Computation Center Duane D. Johnson and Richard M. Martin (PIs) Funded by NSF DMR.
Scheduling Many-Body Short Range MD Simulations on a Cluster of Workstations and Custom VLSI Hardware Sumanth J.V, David R. Swanson and Hong Jiang University.
Parallel Rendering. 2 Introduction In many situations, a standard rendering pipeline might not be sufficient ­Need higher resolution display ­More primitives.
Protein Explorer: A Petaflops Special Purpose Computer System for Molecular Dynamics Simulations David Gobaud Computational Drug Discovery Stanford University.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Parallel & Cluster Computing N-Body Simulation and Collective Communications Henry Neeman, Director OU Supercomputing Center for Education & Research University.
Computing Environment The computing environment rapidly evolving ‑ you need to know not only the methods, but also How and when to apply them, Which computers.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
Parallel Programming & Cluster Computing N-Body Simulation and Collective Communications Henry Neeman, University of Oklahoma Paul Gray, University of.
The Astrophysical MUltiscale Software Environment (AMUSE) P-I: Portegies Zwart Co-Is: Nelemans, Pols, O’Nuallain, Spaans Adv.: Langer, Tolstoy, Hut, Ercolano,
What is QCD? Quantum ChromoDynamics is the theory of the strong force
Gravitational N-body Simulation Major Design Goals -Efficiency -Versatility (ability to use different numerical methods) -Scalability Lesser Design Goals.
Interconnection network network interface and a case study.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Intermediate-mass Black Holes in Star Clusters Holger Baumgardt Astrophysical Computing Laboratory, RIKEN, Tokyo new address:
GRAPE. An implementation of Barnes-Hut treecode on GRAPE-6 with parallel host Makino (2005) For collisionless system Barnes’ vectorization
Computer Operation. Binary Codes CPU operates in binary codes Representation of values in binary codes Instructions to CPU in binary codes Addresses in.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
Dr Andrew Peter Hammersley ESRF ESRF MX COMPUTIONAL AND NETWORK CAPABILITIES (Note: This is not my field, so detailed questions will have to be relayed.
Evolution at CERN E. Da Riva1 CFD team supports CERN development 19 May 2011.
High performance astrophysical simulations using hybrid multiscale and multiphysics methods Simon Portegies Zwart University of Amsterdam Frontiers in.
Setup distribution of N particles
The Earth Simulator System
Super Computing By RIsaj t r S3 ece, roll 50.
Constructing a system with multiple computers or processors
evolution, core collapse, binaries, collisions
BlueGene/L Supercomputer
Physics-based simulation for visual computing applications
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
TeraScale Supernova Initiative
Setup distribution of N particles
P3MSPH-GRAPE Simulations with FIRST Cluster
Types of Parallel Computers
Presentation transcript:

gravitySimulator Beyond the Million Body Problem Collaborators:Rainer Spurzem (Heidelberg) Peter Berczik (Heidelberg/Kiev) Simon Portegies Zwart (Amsterdam) Alessia Gualandris (Amsterdam) Hans-Peter Bischof (RIT) Stefan Harfst and David Merritt Rochester Institute of Technology

Modelling Dense Stellar Systems  one approach: direct N-body simulations  exact but very compute-intensive ~O(N 2 )  many problems require large N  e.g. the evolution of binary Black holes  “empty losscone” is artificially repopulated by two-body scattering unless N > 10 6

How to deal with large N  A standard Supercomputer  Special-purpose hardware  GRAvity PipEline (GRAPE) (J. Makino, T. Fukushige)  Customed-designed pipelines for force calculations  Very fast (~1 TFlops)  Limited particle numbers (< 1/4 million)  Cost: ~$50K + extras (GRAPE-6)

The GRAPE cluster mini-GRAPEs (GRAPE-6A) N < 131,072

GRAPE cluster RIT’s gravitySimulator is operational since Feb 2005   32 dual 3GHz-Xeon nodes   32 GRAPE-6A’s   14 Tbyte RAID   low-latency Infiniband interconnects (10Gbps)   Speed: 4 TFlops   N up to 4 Million particles   Cost: $0.5x10 6   Funding: NSF/NASA/RIT   Next largest:   24 nodes (University of Tokyo)   soon 32 nodes (Heidelberg)

The Code and Performance  new parallel direct-summation code  fourth-order Hermite integrator  individual, block time steps  achieves best performance  for small particle numbers communication dominates  efficiencies are between 60% (many processors) and 90% (few processors) For details see poster GRAPE PC store local particles select active particles collect all active particles compute local force and sum over all nodes

Visualization of N-Body Simulations  new software package “Spiegel”  GUI to plot N-body data and make movies  See Poster for details in collaboration with Hans-Peter Bischof (RIT)