NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.

Slides:



Advertisements
Similar presentations
Severs AIST Cluster (50 CPU) Titech Cluster (200 CPU) KISTI Cluster (25 CPU) Climate Simulation on ApGrid/TeraGrid at SC2003 Client (AIST) Ninf-G Severs.
Advertisements

The Charm++ Programming Model and NAMD Abhinav S Bhatele Department of Computer Science University of Illinois at Urbana-Champaign
1 NAMD - Scalable Molecular Dynamics Gengbin Zheng 9/1/01.
1 Coven a Framework for High Performance Problem Solving Environments Nathan A. DeBardeleben Walter B. Ligon III Sourabh Pandit Dan C. Stanzione Jr. Parallel.
10/21/20091 Protein Explorer: A Petaflops Special-Purpose Computer System for Molecular Dynamics Simulations Makoto Taiji, Tetsu Narumi, Yousuke Ohno,
Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana-Champaign Sameer Kumar IBM T. J. Watson Research Center.
Abhinav Bhatele, Laxmikant V. Kale University of Illinois at Urbana-Champaign Sameer Kumar IBM T. J. Watson Research Center.
City University London
Components for high performance grid programming in the GRID.it project 1 Workshop on Component Models and Systems for Grid Applications - St.Malo 26 june.
Adaptive MPI Chao Huang, Orion Lawlor, L. V. Kalé Parallel Programming Lab Department of Computer Science University of Illinois at Urbana-Champaign.
Tiered architectures 1 to N tiers. 2 An architectural history of computing 1 tier architecture – monolithic Information Systems – Presentation / frontend,
Topology Aware Mapping for Performance Optimization of Science Applications Abhinav S Bhatele Parallel Programming Lab, UIUC.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Client/Server Architectures
A Framework for Collective Personalized Communication Laxmikant V. Kale, Sameer Kumar, Krishnan Varadarajan.
Charm++ Load Balancing Framework Gengbin Zheng Parallel Programming Laboratory Department of Computer Science University of Illinois at.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
BTRC for Macromolecular Modeling and Bioinformatics Beckman Institute, UIUC 1 Future Direction with NAMD David Hardy
1CPSD NSF/DARPA OPAAL Adaptive Parallelization Strategies using Data-driven Objects Laxmikant Kale First Annual Review October 1999, Iowa City.
Alok 1Northwestern University Access Patterns, Metadata, and Performance Alok Choudhary and Wei-Keng Liao Department of ECE,
MSc in High Performance Computing Computational Chemistry Module Parallel Molecular Dynamics (ii) Bill Smith CCLRC Daresbury Laboratory
SOS7, Durango CO, 4-Mar-2003 Scaling to New Heights Retrospective IEEE/ACM SC2002 Conference Baltimore, MD Distilled [Trimmed & Distilled for SOS7 by M.
Designing and Evaluating Parallel Programs Anda Iamnitchi Federated Distributed Systems Fall 2006 Textbook (on line): Designing and Building Parallel Programs.
Scaling to New Heights Retrospective IEEE/ACM SC2002 Conference Baltimore, MD.
Adaptive MPI Milind A. Bhandarkar
1 Scalable Molecular Dynamics for Large Biomolecular Systems Robert Brunner James C Phillips Laxmikant Kale.
DISTRIBUTED COMPUTING
Computational Design of the CCSM Next Generation Coupler Tom Bettge Tony Craig Brian Kauffman National Center for Atmospheric Research Boulder, Colorado.
Young Suk Moon Chair: Dr. Hans-Peter Bischof Reader: Dr. Gregor von Laszewski Observer: Dr. Minseok Kwon 1.
Molecular Dynamics Collection of [charged] atoms, with bonds – Newtonian mechanics – Relatively small #of atoms (100K – 10M) At each time-step – Calculate.
Scheduling Many-Body Short Range MD Simulations on a Cluster of Workstations and Custom VLSI Hardware Sumanth J.V, David R. Swanson and Hong Jiang University.
Programming Models & Runtime Systems Breakout Report MICS PI Meeting, June 27, 2002.
BTRC for Macromolecular Modeling and Bioinformatics Beckman Institute, UIUC 1 Demonstration: Using NAMD David Hardy
Supercomputing Cross-Platform Performance Prediction Using Partial Execution Leo T. Yang Xiaosong Ma* Frank Mueller Department of Computer Science.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
An FPGA Implementation of the Ewald Direct Space and Lennard-Jones Compute Engines By: David Chui Supervisor: Professor P. Chow.
Overcoming Scaling Challenges in Bio-molecular Simulations Abhinav Bhatelé Sameer Kumar Chao Mei James C. Phillips Gengbin Zheng Laxmikant V. Kalé.
Framework for MDO Studies Amitay Isaacs Center for Aerospace System Design and Engineering IIT Bombay.
1CPSD Software Infrastructure for Application Development Laxmikant Kale David Padua Computer Science Department.
Workshop BigSim Large Parallel Machine Simulation Presented by Eric Bohm PPL Charm Workshop 2004.
CCA Common Component Architecture CCA Forum Tutorial Working Group CCA Status and Plans.
1 ©2004 Board of Trustees of the University of Illinois Computer Science Overview Laxmikant (Sanjay) Kale ©
Parallelizing Spacetime Discontinuous Galerkin Methods Jonathan Booth University of Illinois at Urbana/Champaign In conjunction with: L. Kale, R. Haber,
Anton, a Special-Purpose Machine for Molecular Dynamics Simulation By David E. Shaw et al Presented by Bob Koutsoyannis.
Parallelization Strategies Laxmikant Kale. Overview OpenMP Strategies Need for adaptive strategies –Object migration based dynamic load balancing –Minimal.
A uGNI-Based Asynchronous Message- driven Runtime System for Cray Supercomputers with Gemini Interconnect Yanhua Sun, Gengbin Zheng, Laximant(Sanjay) Kale.
Group Mission and Approach To enhance Performance and Productivity in programming complex parallel applications –Performance: scalable to thousands of.
NGS Workshop: Feb 2002PPL-Dept of Computer Science, UIUC Programming Environment and Performance Modeling for million-processor machines Laxmikant (Sanjay)
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
1 Rocket Science using Charm++ at CSAR Orion Sky Lawlor 2003/10/21.
NIH Resource for Macromolecular Modeling and Bioinformatics Beckman Institute, UIUC Scaling NAMD to 100 Million Atoms on Petascale.
Hierarchical Load Balancing for Large Scale Supercomputers Gengbin Zheng Charm++ Workshop 2010 Parallel Programming Lab, UIUC 1Charm++ Workshop 2010.
Parallel Molecular Dynamics A case study : Programming for performance Laxmikant Kale
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
Flexibility and Interoperability in a Parallel MD code Robert Brunner, Laxmikant Kale, Jim Phillips University of Illinois at Urbana-Champaign.
OpenMosix, Open SSI, and LinuxPMI
Parallel Objects: Virtualization & In-Process Components
Performance Evaluation of Adaptive MPI
Component Frameworks:
Milind A. Bhandarkar Adaptive MPI Milind A. Bhandarkar
CSE8380 Parallel and Distributed Processing Presentation
Hybrid Programming with OpenMP and MPI
Department of Computer Science, University of Tennessee, Knoxville
IXPUG, SC’16 Lightning Talk Kavitha Chandrasekar*, Laxmikant V. Kale
An Orchestration Language for Parallel Objects
Support for Adaptivity in ARMCI Using Migratable Objects
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor Dept. of Computer Science

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Vision Make NAMD a widely used MD program –For large molecular systems, –Scaling from PCs, clusters, to large parallel machines –For interactive molecular dynamics Goals: –High performance –Ease of use: configuration and run –Ease of modification (for us and advanced users) Maximize reuse of communication and control patterns Push parallel complexity down into Charm++ runtime –Incorporation of features needed by Scientists

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD 3 New Features Software Goal: –Modular architecture to permit reuse extensibility Scientific/Numeric Modules: –Implicit solvent models (e.g, generalized Born) –Replica exchange (e.g., 10 on 16 processors) –Self-consistent polarizability with a (sequential) CPU penalty of less than 100%. –Hybrid quantum/classical mechanics –Fast nonperiodic (and periodic) electrostatics using multiple grid methods. –A Langevin integrator that permits larger time steps (by being exact for constant forces). –An integrator module that computes shadow energy.

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Design NAMD 3 will be a major rewrite of NAMD –Incorporate lessons learned in the past years –Use modern features of Charm++ –Refactor software for modularity –Restructure for supporting planned features –Algorithms that scale to even larger machines

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Programmability NAMD3 Scientific Modules: –Forces, integration, steering, analysis –Keep code with a common goal together –Add new features without touching old code Parallel Decomposition Framework: –Support common scientific algorithm patterns –Avoid duplicating services for each algorithm –Start with NAMD 2 architecture (but not code)

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Core CHARM++ ClustersLemieuxTeragrid Collective communicationLoad balancer FFTFault ToleranceGrid Scheduling Bonds related Force calculation IntegrationPair-wise Forces calculation PME Charm++ modules NAMD Core Replica exchangeQMImplicit SolventsPolarizable Force Field MDAPI … New Science modules

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC MDAPI Modular Interface Separate “front end” from modular “engine” Same program or over a network or grid Dynamic discovery of engine capabilities, no limitations imposed by interface Front ends: NAMD 2, NAMD 3, Amber, CHARMM, VMD Engines: NAMD 2, NAMD 3, MINDY

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Terascale Biology and Resources PSC LeMieux Riken MDGRAPE NCSA Tungsten TeraGrid ASCI Purple Red Storm Thor’s Hammer CRAY X1

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD on Charm++ Active computer science collaboration (since 1992) Object array - A collection of chares, –with a single global name for the collection, and –each member addressed by an index –Mapping of element objects to processors handled by the system A[0]A[1]A[2]A[3]A[..] A[3]A[0] User’s view System view

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD3 Features Based on Charm++ Adaptive load balancing Optimized communication –Persistent Communication, Optimized concurrent multicast/reduction Flexible, tuned, parallel FFT libraries Automatic Checkpointing Ability to change the number of processors Scheduling on the grid Fault tolerance –Fully automated restart –Survive loss of a node Scaling to large machines –fine-grained parallelism for PME: bonded and nonbonded force evaluations

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Efficient Parallelization for IMD Characteristics –Limited parallelism on small systems –Real time response needed Fine grained parallelization –Improve speedups on 4K-30K atom systems –Time/step goal Currently 0.2s/step for BrH on single processor (P4 1.7GHz) Targeting on 0.003s/step on 64 processors of faster machine, that is 20picosecond/minute Flexible use of clusters –Migrating jobs (shrink/expand) –Better utilization when machine is idle

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Integration with CHARMM/Amber? Goal: NAMD as parallel simulation engine for CHARMM/Amber Generate input files in CHARMM/Amber –NAMD must read native file formats Run with NAMD on parallel computer –Need to use equivalent algorithms Analyze simulation in CHARMM/Amber –NAMD must generate native file formats

NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC Proud of Programmers