ChaNGa CHArm N-body GrAvity. Thomas Quinn Graeme Lufkin Joachim Stadel Laxmikant Kale Filippo Gioachin Pritish Jetley Celso Mendes Amit Sharma.

Slides:



Advertisements
Similar presentations
Parallel Strategies Partitioning consists of the following steps –Divide the problem into parts –Compute each part separately –Merge the results Divide.
Advertisements

Simulating the joint evolution of quasars, galaxies and their large-scale distribution Springel et al., 2005 Presented by Eve LoCastro October 1, 2009.
Course Outline Introduction in algorithms and applications Parallel machines and architectures Overview of parallel machines, trends in top-500 Cluster.
Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana-Champaign Sameer Kumar IBM T. J. Watson Research Center.
Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana-Champaign Sameer Kumar IBM T. J. Watson Research Center.
Astro-2: History of the Universe Lecture 4; April
Numerical issues in SPH simulations of disk galaxy formation Tobias Kaufmann, Lucio Mayer, Ben Moore, Joachim Stadel University of Zürich Institute for.
Non-linear matter power spectrum to 1% accuracy between dynamical dark energy models Matt Francis University of Sydney Geraint Lewis (University of Sydney)
PRESENTATION TOPIC  DARK MATTER &DARK ENERGY.  We know about only normal matter which is only 5% of the composition of universe and the rest is  DARK.
ASCI/Alliances Center for Astrophysical Thermonuclear Flashes Simulating Self-Gravitating Flows with FLASH P. M. Ricker, K. Olson, and F. X. Timmes Motivation:
Module on Computational Astrophysics Professor Jim Stone Department of Astrophysical Sciences and PACM.
In The Beginning  N-body simulations (~100s particles) – to study Cluster formation  Cold collapse produces too steep a density profile (Peebles 1970)
Universe in a box: simulating formation of cosmic structures Andrey Kravtsov Department of Astronomy & Astrophysics Center for Cosmological Physics (CfCP)
Daniel Blackburn Load Balancing in Distributed N-Body Simulations.
1 Parallel multi-grid summation for the N-body problem Jesús A. Izaguirre with Thierry Matthey Department of Computer Science and Engineering University.
Derek C. Richardson (U Maryland) PKDGRAV : A Parallel k-D Tree Gravity Solver for N-Body Problems FMM 2004.
Cosmological N-body simulations of structure formation Jürg Diemand, Ben Moore and Joachim Stadel, University of Zurich.
PRESIDENCY UNIVERSITY
박창범 ( 고등과학원 ) & 김주한 ( 경희대학교 ), J. R. Gott (Princeton, USA), J. Dubinski (CITA, Canada) 한국계산과학공학회 창립학술대회 Cosmological N-Body Simulation of Cosmic.
Topology Aware Mapping for Performance Optimization of Science Applications Abhinav S Bhatele Parallel Programming Lab, UIUC.
N Tropy: A Framework for Analyzing Massive Astrophysical Datasets Harnessing the Power of Parallel Grid Resources for Astrophysical Data Analysis Jeffrey.
Parallel Performance of Hierarchical Multipole Algorithms for Inductance Extraction Ananth Grama, Purdue University Vivek Sarin, Texas A&M University Hemant.
The Evolution of the Universe Nicola Loaring. The Big Bang According to scientists the Universe began ~15 billion years ago in a hot Big Bang. At creation.
Course Outline Introduction in algorithms and applications Parallel machines and architectures Overview of parallel machines, trends in top-500, clusters,
INAF Osservatorio Astrofisico di Catania “ScicomP 9” Bologna March 23 – Using LAPI and MPI-2 in an N-body cosmological code on IBM SP in an N-body.
Cosmology I & II Fall 2012 Cosmology Cosmology I & II  Cosmology I:  Cosmology II: 
Intro to Cosmology! OR What is our Universe?. The Latest High Resolution Image of the Cosmic Microwave Background Radiation Low Energy RegionHigh Energy.
Cosmology and extragalactic astronomy
Origin of solar systems 30 June - 2 July 2009 by Klaus Jockers Max-Planck-Institut of Solar System Science Katlenburg-Lindau.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Sean Passmoor Supervised by Dr C. Cress Simulating the Radio Sky.
PARALLELIZING HIGHLY DYNAMIC N-BODY SIMULATIONS Joachim Stadel University of Zurich Institute for Theoretical Physics.
Renaissance: Formation of the first light sources in the Universe after the Dark Ages Justin Vandenbroucke, UC Berkeley Physics 290H, February 12, 2008.
Slides for Parallel Programming Techniques & Applications Using Networked Workstations & Parallel Computers 2nd ed., by B. Wilkinson & M
Racah Institute of physics, Hebrew University (Jerusalem, Israel)
Scientific Motivations for the VO Historical Remarks on Massive Data Collection Projects. (Obvious) Potential Virtues and Defects of the VO for Observations.
Cosmology and Dark Matter III: The Formation of Galaxies Jerry Sellwood.
Scalable and Topology-Aware Load Balancers in Charm++ Amit Sharma Parallel Programming Lab, UIUC.
Energy Balance in Clusters of Galaxies Patrick M. Motl & Jack O. Burns Center for Astrophysics and Space Astronomy University of Colorado at Boulder X-ray.
What the Formation of the First Stars Left in its Wake.
SPH Simulations of the Galaxy Evolution NAKASATO, Naohito University of Tokyo.
Improved Chemical Evolution Model for the Early Galaxy Evolution 中里直人 東大天文.
Hierarchical Load Balancing for Large Scale Supercomputers Gengbin Zheng Charm++ Workshop 2010 Parallel Programming Lab, UIUC 1Charm++ Workshop 2010.
Study of Proto-clusters by Cosmological Simulation Tamon SUWA, Asao HABE (Hokkaido Univ.) Kohji YOSHIKAWA (Tokyo Univ.)
Teragrid 2009 Scalable Interaction with Parallel Applications Filippo Gioachin Chee Wai Lee Laxmikant V. Kalé Department of Computer Science University.
ChaNGa CHArm N-body GrAvity. Thomas Quinn Graeme Lufkin Joachim Stadel James Wadsley Greg Stinson Laxmikant Kale Filippo Gioachin Pritish Jetley Celso.
A novel approach to visualizing dark matter simulations
Massively Parallel Cosmological Simulations with ChaNGa Pritish Jetley, Filippo Gioachin, Celso Mendes, Laxmikant V. Kale and Thomas Quinn.
1 Scalable Cosmological Simulations on Parallel Machines Filippo Gioachin¹ Amit Sharma¹ Sayantan Chakravorty¹ Celso Mendes¹ Laxmikant V. Kale¹ Thomas R.
ChaNGa CHArm N-body GrAvity. Thomas Quinn Graeme Lufkin Joachim Stadel James Wadsley Laxmikant Kale Filippo Gioachin Pritish Jetley Celso Mendes Amit.
Cosmology On Petascale Computers. Fabio Governato, Greg Stinson, Chris Brook, Alison Brooks Joachim Stadel, Lucio Mayer, Ben Moore, George Lake (U. ZH)
1 ChaNGa: The Charm N-Body GrAvity Solver Filippo Gioachin¹ Pritish Jetley¹ Celso Mendes¹ Laxmikant Kale¹ Thomas Quinn² ¹ University of Illinois at Urbana-Champaign.
ChaNGa CHArm N-body GrAvity. Thomas Quinn Graeme Lufkin Joachim Stadel Laxmikant Kale Filippo Gioachin Pritish Jetley Celso Mendes Amit Sharma.
Progressive Clustering of Big Data with GPU Acceleration and Visualization Jun Wang1, Eric Papenhausen1, Bing Wang1, Sungsoo Ha1, Alla Zelenyuk2, and Klaus.
Hydrodynamic Galactic Simulations
ChaNGa: Design Issues in High Performance Cosmology
Outline Part II. Structure Formation: Dark Matter
Tamas Szalay, Volker Springel, Gerard Lemson
Performance Evaluation of Adaptive MPI
Implementing Simplified Molecular Dynamics Simulation in Different Parallel Paradigms Chao Mei April 27th, 2006 CS498LVK.
dark matter and the Fate of the Universe
Course Outline Introduction in algorithms and applications
Radiation Hydrodynamics with HMCS
Radiation Hydrodynamics with HMCS
Case Studies with Projections
Formation of the Universe
Outline Part II. Structure Formation: Dark Matter
Gengbin Zheng, Esteban Meneses, Abhinav Bhatele and Laxmikant V. Kale
Support for Adaptivity in ARMCI Using Migratable Objects
N-Body Gravitational Simulations
Presentation transcript:

ChaNGa CHArm N-body GrAvity

Thomas Quinn Graeme Lufkin Joachim Stadel Laxmikant Kale Filippo Gioachin Pritish Jetley Celso Mendes Amit Sharma

Outline ● Scientific background – How to build a Galaxy – Types of Simulations – Simulation Challenges ● ChaNGa and those Challenges – Features – Tree gravity – Load balancing – Multistepping ● Future Challenges – Needed Simulations – Technology Challenges

Image courtesy NASA/WMAP Cosmology: How does this...

... turn into this?

Computational Cosmology ● CMB gives fluctuations of 1e-5 ● Galaxies are overdense by 1e7 ● It happens through Gravitational Collapse ● Making testable predictions from a cosmological hypothesis requires – Non-linear, dynamic calculation – e.g. Computer simulation

Simulation process ● Start with fluctuations based on Dark Matter properties ● Follow model analytically (good enough to get CMB) ● Create a realization of these fluctuations in particles. ● Follow the motions of these particles as they interact via gravity. ● Compare final distribution of particles with observed properties of galaxies.

Simulating galaxies: Procedure 1. Simulate 100 Mpc volume at kpc resolution 2. Pick candidate galaxies for further study 3. Resimulate galaxies with same large scale structure but with higher resolution, and lower resolution in the rest of the computational volume. 4. At higher resolutions, include gas physics and star formation.

Gas Stars Dark Matter

9/30/2016Parallel Programming UIUC11 Types of simulations Zoom In “Uniform” Volume Star Cluster

Computational Challenges ● Large spacial dynamic range: > 100 Mpc to < 1 kpc – Hierarchical, adaptive gravity solver is needed ● Large temporal dynamic range: 10 Gyr to 1 Myr – Multiple timestep algorithm is needed ● Gravity is a long range force – Hierarchal information needs to go across processor domains

● Multi - Platform ● Massively Parallel (100s; 1000s on large sims) ● Treecode with periodic boundary conditions ● Multi-stepping (but bad load balancing) ● Hydrodynamics (via SPH) with radiative cooling ● UV background ● Star Formation ● Supernovae feedback into thermal energy The existing code:

ChaNGa Features ● Tree-based gravity solver ● High order multipole expansion ● Periodic boundaries (if needed) ● Individual multiple timesteps ● Dynamic load balancing with choice of strategies ● Checkpointing ● Visualization ● Built from the ground up on Charm++

Need for high multipole order

16 Parallel Programming UIUC 9/30/2016 Space decomposition TreePiece 1TreePiece 2TreePiece 3...

17 Parallel Programming UIUC 9/30/2016 Basic algorithm... ● Newtonian gravity interaction – Each particle is influenced by all others: O(n²) algorithm ● Barnes-Hut approximation: O(nlogn) – Influence from distant particles combined into center of mass

18 Parallel Programming UIUC 9/30/ in parallel ● Remote data – need to fetch from other processors ● Data reusage – same data needed by more than one particle

19 Parallel Programming UIUC 9/30/2016 Overall algorithm Processor 1 local work (low priority) remote work mi ss TreePiece C local work (low priority) remote work TreePiece B global work prefetc h visit of the tree TreePiece A local work (low priority) Start computation End computation global work re mo te present? request node CacheManager YES: return Processor n reply with requested data NO: fetch callback TreePiece on Processor 2 buff er High priority prefetc h visit of the tree

9/30/2016Parallel Programming UIUC20 Scaling: comparison Uniform 3M on Tungsten

9/30/2016Parallel Programming UIUC21 Load balancing with GreedyLB Zoom In 5M on 1,024 BlueGene/L processors 5.6s6.1s 4x messages

9/30/2016Parallel Programming UIUC22 Load balancing with OrbRefineLB Zoom in 5M on 1,024 BlueGene/L processors 5.6s 5.0s

9/30/2016Parallel Programming UIUC23 Scaling with load balancing Number of Processors x Execution Time per Iteration (s)

Timestepping Challenges ● 1/m particles need m times more force evaluations ● Naively, simulation cost scales as N^(4/3)ln(N) – This is a problem when N ~ 1e9 or greater ● If each particle an individual timestep scaling reduces to N (ln(N))^2 ● A difficult dynamic load balancing problem

Timestepping and Load Balancing

Cosmo Loadbalancer ● Use Charm++ measurement based load balancer ● Modification: provide LB database with information about timestepping. – “Large timestep”: balance based on previous Large step – “Small step” balance based on previous small step

Results on 3 rung example 613s 429s 228s

Summary ● Cosmological simulations provide a challenges to parallel implementations – Non-local data dependencies – Hierarchical in space and time ● ChaNGa has been successful in addressing this challenges using Charm++ features – Message priorities – New load balancers

Future ● Changa currently in use in high time dynamic range simulations: galactic nuclei ● New Physics – Smooth particle hydrodynamics ● Better gravity algorithms – Fast multipole method – New domain decomposition/load balancing strategies ● Generic tree walk to enable new algorithms

Have We converged? Weinberg & Katz (2007)

Computing Challenge Summary ● The Universe is big => we will always be pushing for more resources ● New algorithm efforts will be made to make efficient use of the resources we have – Efforts made to abstract away from machine details – Parallelization efforts need to depend on more automated processes.