PERFORMANCE EVALUATION OF USER-REAXC PACKAGE Hasan Metin Aktulga Postdoctoral Researcher Scientific Computing Group Lawrence Berkeley National Laboratory.

Slides:



Advertisements
Similar presentations
Making Time-stepped Applications Tick in the Cloud Tao Zou, Guozhang Wang, Marcos Vaz Salles*, David Bindel, Alan Demers, Johannes Gehrke, Walker White.
Advertisements

Parallel Jacobi Algorithm Steven Dong Applied Mathematics.
Parameterizing a Geometry using the COMSOL Moving Mesh Feature
PHYS466 Project Kyoungmin Min, Namjung Kim and Ravi Bhadauria.
Formulation of an algorithm to implement Lowe-Andersen thermostat in parallel molecular simulation package, LAMMPS Prathyusha K. R. and P. B. Sunil Kumar.
OpenFOAM on a GPU-based Heterogeneous Cluster
Hierarchical Multi-Resolution Finite Element Model for Soft Body Simulation Matthieu Nesme, François Faure, Yohan Payan 2 nd Workshop on Computer Assisted.
CISC October Goals for today: Foster’s parallel algorithm design –Partitioning –Task dependency graph Granularity Concurrency Collective communication.
PRISM Mid-Year Review Reactive Atomistics, Molecular Dynamics.
Molecular Dynamics Molecular dynamics Some random notes on molecular dynamics simulations Seminar based on work by Bert de Groot and many anonymous Googelable.
Avoiding Communication in Sparse Iterative Solvers Erin Carson Nick Knight CS294, Fall 2011.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
Monica Garika Chandana Guduru. METHODS TO SOLVE LINEAR SYSTEMS Direct methods Gaussian elimination method LU method for factorization Simplex method of.
PuReMD: A Reactive (ReaxFF) Molecular Dynamics Package Hassan Metin Akgulta, Sagar Pandit, Joseph Fogarty, Ananth Grama Acknowledgements:
Efficient Parallel Implementation of Molecular Dynamics with Embedded Atom Method on Multi-core Platforms Reporter: Jilin Zhang Authors:Changjun Hu, Yali.
MUMPS A Multifrontal Massively Parallel Solver IMPLEMENTATION Distributed multifrontal.
Parallel Performance of Hierarchical Multipole Algorithms for Inductance Extraction Ananth Grama, Purdue University Vivek Sarin, Texas A&M University Hemant.
An approach for solving the Helmholtz Equation on heterogeneous platforms An approach for solving the Helmholtz Equation on heterogeneous platforms G.
Algorithms and Software for Large-Scale Simulation of Reactive Systems _______________________________ Ananth Grama Coordinated Systems Lab Purdue University.
Next-generation scalable applications: When MPI-only is not enough June 3, 2008 Paul S. Crozier Sandia National Labs Sandia is a multiprogram laboratory.
Stratified Magnetohydrodynamics Accelerated Using GPUs:SMAUG.
1 A Domain Decomposition Analysis of a Nonlinear Magnetostatic Problem with 100 Million Degrees of Freedom H.KANAYAMA *, M.Ogino *, S.Sugimoto ** and J.Zhao.
High-Performance Quantum Simulation: A challenge to Schr ö dinger equation on 256^4 grids * Toshiyuki Imamura 13 今村俊幸, Thanks to Susumu Yamada 23, Takuma.
PuReMD: Purdue Reactive Molecular Dynamics Package Hasan Metin Aktulga and Ananth Grama Purdue University TST Meeting,May 13-14, 2010.
Reactive Molecular Dynamics: Progress Report Hassan Metin Aktulga 1, Joseph Fogarty 2, Sagar Pandit 2, and Ananth Grama 3 1 Lawrence Berkeley Lab 2 University.
Algorithmic and Numerical Techniques for Atomistic Modeling Hasan Metin Aktulga PhD Thesis Defense June 23 rd, 2010.
Presented by High Productivity Language and Systems: Next Generation Petascale Programming Wael R. Elwasif, David E. Bernholdt, and Robert J. Harrison.
ParCFD Parallel computation of pollutant dispersion in industrial sites Julien Montagnier Marc Buffat David Guibert.
PuReMD: A Reactive (ReaxFF) Molecular Dynamics Package Ananth Grama and Metin Aktulga
Scalable Symbolic Model Order Reduction Yiyu Shi*, Lei He* and C. J. Richard Shi + *Electrical Engineering Department, UCLA + Electrical Engineering Department,
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Scaling Area Under a Curve. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
1 1  Capabilities: Dynamic load balancing and static data partitioning -Geometric, graph-based, hypergraph-based -Interfaces to ParMETIS, PT-Scotch, PaToH.
Overcoming Scaling Challenges in Bio-molecular Simulations Abhinav Bhatelé Sameer Kumar Chao Mei James C. Phillips Gengbin Zheng Laxmikant V. Kalé.
Molecular simulation methods Ab-initio methods (Few approximations but slow) DFT CPMD Electron and nuclei treated explicitly. Classical atomistic methods.
Algorithms, Numerical Techniques, and Software for Atomistic Modeling.
Molecular Simulation of Reactive Systems. _______________________________ Sagar Pandit, Hasan Aktulga, Ananth Grama Coordinated Systems Lab Purdue University.
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
Algorithms and Software for Large-Scale Simulation of Reactive Systems _______________________________ Metin Aktulga, Sagar Pandit, Alejandro Strachan,
Reactive Molecular Dynamics: Algorithms, Software, and Applications. Ananth Grama Computer Science, Purdue University
Parallel Solution of the Poisson Problem Using MPI
Enabling Self-management of Component-based High-performance Scientific Applications Hua (Maria) Liu and Manish Parashar The Applied Software Systems Laboratory.
Finite Element Modelling of Photonic Crystals Ben Hiett J Generowicz, M Molinari, D Beckett, KS Thomas, GJ Parker and SJ Cox High Performance Computing.
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
Scaling Conway’s Game of Life. Why do parallelism? Speedup – solve a problem faster. Accuracy – solve a problem better. Scaling – solve a bigger problem.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
PuReMD Design Initialization – neighbor-list, bond-list, hydrogenbond-list and Coefficients of QEq matrix Bonded interactions – Bond-order, bond-energy,
A Parallel Hierarchical Solver for the Poisson Equation Seung Lee Deparment of Mechanical Engineering
Multipole-Based Preconditioners for Sparse Linear Systems. Ananth Grama Purdue University. Supported by the National Science Foundation.
Parallel Molecular Dynamics A case study : Programming for performance Laxmikant Kale
PuReMD: Purdue Reactive Molecular Dynamics Software Ananth Grama Center for Science of Information Computational Science and Engineering Department of.
Accelerating K-Means Clustering with Parallel Implementations and GPU Computing Janki Bhimani Miriam Leeser Ningfang Mi
1 Nanoscale Modeling and Computational Infrastructure ___________________________ Ananth Grama Professor of Computer Science, Associate Director, PRISM.
1 Des bulles dans le PaStiX Réunion NUMASIS Mathieu Faverge ScAlApplix project, INRIA Futurs Bordeaux 29 novembre 2006.
Block Low Rank Approximations in LS-DYNA
Parallel Programming By J. H. Wang May 2, 2017.
Implementing Simplified Molecular Dynamics Simulation in Different Parallel Paradigms Chao Mei April 27th, 2006 CS498LVK.
Parallel Programming in C with MPI and OpenMP
A Parallel Hierarchical Solver for the Poisson Equation
David Gleich, Ahmed Sameh, Ananth Grama, et al.
Comparison to LAMMPS-REAX
Algorithms and Software for Large-Scale Simulation of Reactive Systems
PuReMD: Purdue Reactive Molecular Dynamics Software Ananth Grama Center for Science of Information Computational Science and Engineering Department of.
Supported by the National Science Foundation.
Mid-Year Review Template March 2, 2010 Purdue University
Molecular simulation methods
Algorithms and Software for Large-Scale Simulation of Reactive Systems
Ph.D. Thesis Numerical Solution of PDEs and Their Object-oriented Parallel Implementations Xing Cai October 26, 1998.
Parallel Programming in C with MPI and OpenMP
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

PERFORMANCE EVALUATION OF USER-REAXC PACKAGE Hasan Metin Aktulga Postdoctoral Researcher Scientific Computing Group Lawrence Berkeley National Laboratory LAMMPS Users’ Workshop Aug. 10, 2011, Albuquerque NM

ReaxFF vs. Classical MD Basis: atoms Parameterized interactions Newtonian mechanics Several common interactions ReaxFF Classical MD Fully atomistic Dynamic bonds Dynamic charges Can study reactive systems Sub-femtosecond time-steps May represent atoms in groups Static bonds Static charges (in general) Only non-reactive systems Femtoseconds long time-steps

USER-REAXC Package IS BASED ON… PuReMD (Purdue Reactive Molecular Dynamics) Built on sPuReMD using 3D Mesh decomposition Extends its capabilities to large systems, longer time-scales Demonstrated scalability over 3000 processors GENERAL FEATURES Implemented from scratch in C Modular implementation for further improvements, extensions Distributed freely under GPL RELATED PUBLICATION Parallel Reactive Molecular Dynamics: Numerical Methods and Algorithmic Techniques H. M. Aktulga, J. C. Fogarty, S. A. Pandit, A. Y. Grama To Appear in the Parallel Computing Journal

Tabulated Long-Range Interactions Natural cubic splines for excellent accuracy, low memory footprint Observed speed-ups of 3x USER-REAXC: Key Contributions Advanced Dynamic Memory Management Compact data structures Dynamic and adaptive lists Low memory footprint, linear scaling with system size! Elimination of Bond Order Derivatives List

FIX QEQ/REAX Preconditioned Conjugate Gradient (PCG) solver Diagonal pre-conditioner Smart initial guesses (through linear and quadratic extrapolations) Redundant computations to eliminate a communication step Iterate both systems together to save on communication (TO DO) USER-REAXC: QEq Solver Charge Equilibration (QEq) Problem  Need to solve the following equations: 

PuReMD (USER-REAXC) Scaling Weak Scaling: Bulk Water Per core: 6540 atoms in a 40x40x40 A 3 box Strong Scaling: Bulk Water atoms in a 80x80x80 A 3 box

PuReMD (USER-REAXC) vs. REAX Weak Scaling: Bulk Water Per core: 6540 atoms in a 40x40x40 A 3 box Strong Scaling: Bulk Water atoms in a 80x80x80 A 3 box

On Memory Footprints LAMMPS-REAX: static allocations during compile time USER-REAXC: dynamic and adaptive allocations at runtime Memory Footprints of LAMMPS-REAX vs sPuReMD* * PuReMD: slightly larger memory footprint than sPuReMD * USER-REAXC: extra storage space for LAMMPS internals

Future Work Improve the QEq Solver Improve the SpMV in the iterative solve Use a Neutral-Territory (NT) Method to reduce communication Maybe use communication-avoiding solvers (like CA-GMRES + CA-ILUT) THANK YOU! QUESTIONS? Hybrid OpenMP/MPI implementation to reduce communication Especially necessary as we move towards many-core era