SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Overview Thomas E. Cheatham III University of Utah Jan 14th 2010 By Ross C. Walker.

Slides:



Advertisements
Similar presentations
A very brief introduction to ParaView Jan Götz Erlangen 2007 Department of Computer Science 10 System Simulation.
Advertisements

1 Use Cases Application provisioning (version control) Workload management/load-balancing (server consolidation) Data Federation/sharing E-utilities (provisioning.
Tesla CUDA GPU Accelerates Computing The Right Processor for the Right Task.
MPI version of the Serial Code With One-Dimensional Decomposition Presented by Timothy H. Kaiser, Ph.D. San Diego Supercomputer Center Presented by Timothy.
University of Chicago Department of Energy The Parallel and Grid I/O Perspective MPI, MPI-IO, NetCDF, and HDF5 are in common use Multi TB datasets also.
A PLFS Plugin for HDF5 for Improved I/O Performance and Analysis Kshitij Mehta 1, John Bent 2, Aaron Torres 3, Gary Grider 3, Edgar Gabriel 1 1 University.
Formulation of an algorithm to implement Lowe-Andersen thermostat in parallel molecular simulation package, LAMMPS Prathyusha K. R. and P. B. Sunil Kumar.
NewsFlash!! Earth Simulator no longer #1. In slightly less earthshaking news… Homework #1 due date postponed to 10/11.
SAN DIEGO SUPERCOMPUTER CENTER Blue Gene for Protein Structure Prediction (Predicting CASP Targets in Record Time) Ross C. Walker.
SAN DIEGO SUPERCOMPUTER CENTER HDF5/SRB Integration July 10, 2006 Mike Wan SRB, SDSC Peter Cao
Analyzing MD Results. Extracting system properties from the data written to the mdout files MD information is written in the form: NSTEP = 100 TIME(PS)
SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Outline October 9th 2008 Ross C. Walker.
XML Parsing Using Java APIs AIP Independence project Fall 2010.
Chapter 101 Virtual Memory Chapter 10 Sections and plus (Skip:10.3.2, 10.7, rest of 10.8)
NPACI: National Partnership for Advanced Computational Infrastructure Supercomputing ‘98 Mannheim CRAY T90 vs. Tera MTA: The Old Champ Faces a New Challenger.
Where Are They Now? Current Status of C++ Parallel Language Extensions and Libraries From 1995 Supercomputing Workshop 1.
1 CS 501 Spring 2005 CS 501: Software Engineering Lecture 22 Performance of Computer Systems.
Introduction to Symmetric Multiprocessors Süha TUNA Bilişim Enstitüsü UHeM Yaz Çalıştayı
Virtual Memory.
Ana Damjanovic (JHU, NIH) JHU: Petar Maksimovic Bertrand Garcia-Moreno NIH: Tim Miller Bernard Brooks OSG: Torre Wenaus and team.
BTRC for Macromolecular Modeling and Bioinformatics Beckman Institute, UIUC 1 Future Direction with NAMD David Hardy
Statistical Performance Analysis for Scientific Applications Presentation at the XSEDE14 Conference Atlanta, GA Fei Xing Haihang You Charng-Da Lu July.
SAN DIEGO SUPERCOMPUTER CENTER HDF5/SRB Integration August 28, 2006 Mike Wan SRB, SDSC Peter Cao
BTRC for Macromolecular Modeling and Bioinformatics Beckman Institute, UIUC 1 Demonstration: Using NAMD David Hardy

IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Chapter 9: Virtual Memory Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped Files Allocating Kernel.
SciDAC All Hands Meeting, March 2-3, 2005 Northwestern University PIs:Alok Choudhary, Wei-keng Liao Graduate Students:Avery Ching, Kenin Coloma, Jianwei.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
Parallel I/O Performance: From Events to Ensembles Andrew Uselton National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory.
ALGORITHM IMPROVEMENTS AND HPC 1. Basic MD algorithm 1. Initialize atoms’ status and the LJ potential table; set parameters controlling the simulation;
SAN DIEGO SUPERCOMPUTER CENTER HDF5/SRB Integration June 14, 2006 Mike Wan SRB, SDSC Peter Cao
_______________________________________________________________CMAQ Libraries and Utilities ___________________________________________________Community.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Lecture 16: Reconfigurable Computing Applications November 3, 2004 ECE 697F Reconfigurable Computing Lecture 16 Reconfigurable Computing Applications.
Co-ordination & Harmonisation of Advanced e-Infrastructures for Research and Education Data Sharing Research Infrastructures – Proposal n GROMACs.
Bioinformatics: Practical Application of Simulation and Data Mining Markov Modeling II Prof. Corey O’Hern Department of Mechanical Engineering Department.
Overcoming Scaling Challenges in Bio-molecular Simulations Abhinav Bhatelé Sameer Kumar Chao Mei James C. Phillips Gengbin Zheng Laxmikant V. Kalé.
1 The Performance Analysis of Molecular dynamics RAD GTPase with AMBER application on Cluster computing environtment. The Performance Analysis of Molecular.
1/20 Study of Highly Accurate and Fast Protein-Ligand Docking Method Based on Molecular Dynamics Reporter: Yu Lun Kuo
Comparative Study of NAMD and GROMACS
Molecular Modelling - Lecture 2 Techniques for Conformational Sampling Uses CHARMM force field Written in C++
SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Overview Adrian E. Roitberg University of Florida July 2nd 2009 By Ross C. Walker.
Towards Scalable Performance Analysis and Visualization through Data Reduction Chee Wai Lee, Celso Mendes, L. V. Kale University of Illinois at Urbana-Champaign.
Efficient SAS programming with Large Data Aidan McDermott Computing Group, March 2007.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Advanced User Support for MPCUGLES code at University of Minnesota October 09,
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
SDM Center Parallel I/O Storage Efficient Access Team.
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
Data-Driven Time-Parallelization in the AFM Simulation of Proteins L. Ji, H. Nymeyer, A. Srinivasan, and Y. Yu Florida State University
Virtual Observatory India VOStat Statistical Analysis for the Virtual Observatory By Deoyani and Mohasin.
OCR A Level F453: The function and purpose of translators Translators a. describe the need for, and use of, translators to convert source code.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
SCEC Capability Simulations on TeraGrid
GROMACs GROningen MAchine for Chemical Simulations Eliana Bohorquez Puentes Nicolas Ortiz Gonzalez.
VisIt Project Overview
Development Environment
MASS Java Documentation, Verification, and Testing
Integrating Scientific Tools and Web Portals
Virtual Memory CSSE 332 Operating Systems
Parastou Sadatmousavi§, & Ross C. Walker*
Chapter 9: Virtual Memory
for the Offline and Computing groups
Architecture & System Overview
Cloud Distributed Computing Environment Hadoop
Storage Structure and Efficient File Access
Insight into the stabilization of A-DNA by specific ion association: spontaneous B-DNA to A-DNA transitions observed in molecular dynamics simulations.
Hierarchical Data Format (HDF) Status Update
The site to download BALBES:
Presentation transcript:

SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Overview Thomas E. Cheatham III University of Utah Jan 14th 2010 By Ross C. Walker

SAN DIEGO SUPERCOMPUTER CENTER Project Overview The Problem: Extensive work has been undertaken to improve the performance and scalability of Molecular Dynamics (MD) packages. NAMD AMBER GROMACS LAMMPS DESMOND

SAN DIEGO SUPERCOMPUTER CENTER Project Overview The Problem: Larger and Larger systems are being simulated. (1 million + atoms now routine) Longer and longer simulations are being run. (100ns to 1 microsecond now routine) Use of GPUs is leading to longer simulations being run on ‘basic’ desktops. This is leading to an explosion in trajectory size. The analysis tools are not keeping up with this.

SAN DIEGO SUPERCOMPUTER CENTER Project Overview Development of Parallel Analysis Tools for Molecular Dynamics Project composed of 3 sections: Development of parallel analysis tools for MD. Improvements in the handling of large datasets. Development of new analysis techniques for large and/or long timescale simulations.

SAN DIEGO SUPERCOMPUTER CENTER Phase 1 Parallelization of Ptraj Analysis Suite

SAN DIEGO SUPERCOMPUTER CENTER Ptraj Developed by Prof. Cheatham during the 1990’s at UCSF. Has been continually updated since. Distributed as openSource in the AMBERTools package. Supports a large range of analysis techniques on multiple file formats. (AMBER, CHARMM, NAMD etc…)

SAN DIEGO SUPERCOMPUTER CENTER Work to Date I have written a parallel version of ptraj that uses MPI calls and MPI/IO. It supports parallel I/O to take advantage of Lustre etc. Version 1.0 of the parallel version released at the end of December 2009 as part of free AMBERTools 1.3. (See Multiple performance improvements also included. Random file access. Non-sequential access to frames in trajectory files etc.

SAN DIEGO SUPERCOMPUTER CENTER Version 1.0 Current supported parallel actions are: angle, atomicfluct, average, center, checkoverlap, closest, contacts,dihedrals, distance, image, principal, pucker, radgyr, randomizeions, rms,strip, translate, watershell For actions that are not yet parallel (e.g. hbond) the code will default back to single cpu operation.

SAN DIEGO SUPERCOMPUTER CENTER Example Performance

SAN DIEGO SUPERCOMPUTER CENTER Example Performance

SAN DIEGO SUPERCOMPUTER CENTER Random File Access Improvements The code can now seek randomly even for ASCII trajectory files. trajin foo.mdcrd trajin foo2.mdcrd image origin center trajout foobar.pdb pdb 1 TB trajectory file (Ranger) Original Code:5820 secs (97 mins) New Code:1 cpu37.5 secs 8 cpu4.8 secs

SAN DIEGO SUPERCOMPUTER CENTER Current Issues Cannot currently do random / parallel IO on compressed files. Code currently decompresses the file by calling gzip etc and streaming it. Suggestions? Compressed NetCDF allows this? Use of non-parallel action currently forces the entire run to be serial.

SAN DIEGO SUPERCOMPUTER CENTER Planned Work for the Coming Year Complete parallelization of remaining actions. Develop distributed memory versions of the memory hungry routines (e.g. hbond) to support larger systems. Begin developing (in collaboration with Prof Cheatham’s lab) new analysis techniques that are more useful when looking at microsecond+ trajectories or multi-million atom simulations.