The ACTS Toolkit (What can it do for you?) Osni Marques and Tony Drummond ( LBNL/NERSC )

Slides:



Advertisements
Similar presentations
Advanced Computational Software Scientific Libraries: Part 2 Blue Waters Undergraduate Petascale Education Program May 29 – June
Advertisements

DOE 2000: The Advanced Computational Testing and Simulation Toolkit James R. McGraw Lawrence Livermore National Laboratory.
1 A Common Application Platform (CAP) for SURAgrid -Mahantesh Halappanavar, John-Paul Robinson, Enis Afgane, Mary Fran Yafchalk and Purushotham Bangalore.
Problem-Solving Environments: The Next Level in Software Integration David W. Walker Cardiff University.
Symmetric Eigensolvers in Sca/LAPACK Osni Marques
Reconfigurable Application Specific Computers RASCs Advanced Architectures with Multiple Processors and Field Programmable Gate Arrays FPGAs Computational.
1 NERSC User Group Business Meeting, June 3, 2002 High Performance Computing Research Juan Meza Department Head High Performance Computing Research.
Where Are They Now? Current Status of C++ Parallel Language Extensions and Libraries From 1995 Supercomputing Workshop 1.
Programming Tools and Environments: Linear Algebra James Demmel Mathematics and EECS UC Berkeley.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium Siu Man Yau, Katherine.
PETSc Portable, Extensible Toolkit for Scientific computing.
Asst.Prof.Dr.Ahmet Ünveren SPRING Computer Engineering Department Asst.Prof.Dr.Ahmet Ünveren SPRING Computer Engineering Department.
1 Parallel Simulations of Underground Flow in Porous and Fractured Media H. Mustapha 1,2, A. Beaudoin 1, J. Erhel 1 and J.R. De Dreuzy IRISA – INRIA.
Iterative computation is a kernel function to many data mining and data analysis algorithms. Missing in current MapReduce frameworks is collective communication,
Center for Component Technology for Terascale Simulation Software (aka Common Component Architecture) (aka CCA) Rob Armstrong & the CCA Working Group Sandia.
MUMPS A Multifrontal Massively Parallel Solver IMPLEMENTATION Distributed multifrontal.
Global Address Space Applications Kathy Yelick NERSC/LBNL and U.C. Berkeley.
CS 591x – Cluster Computing and Programming Parallel Computers Parallel Libraries.
S.S. Yang and J.K. Lee FEMLAB and its applications POSTEC H Plasma Application Modeling Lab. Oct. 25, 2005.
Center for Programming Models for Scalable Parallel Computing: Project Meeting Report Libraries, Languages, and Execution Models for Terascale Applications.
1 Using the PETSc Parallel Software library in Developing MPP Software for Calculating Exact Cumulative Reaction Probabilities for Large Systems (M. Minkoff.
Overview of the ACTS Toolkit For NERSC Users Brent Milne John Wu
BLU-ICE and the Distributed Control System Constraints for Software Development Strategies Timothy M. McPhillips Stanford Synchrotron Radiation Laboratory.
ANS 1998 Winter Meeting DOE 2000 Numerics Capabilities 1 Barry Smith Argonne National Laboratory DOE 2000 Numerics Capability
Programming Models & Runtime Systems Breakout Report MICS PI Meeting, June 27, 2002.
1 Cactus in a nutshell... n Cactus facilitates parallel code design, it enables platform independent computations and encourages collaborative code development.
Khoros Yongqun He Dept. of Computer Science, Virginia Tech.
1 Computer Programming (ECGD2102 ) Using MATLAB Instructor: Eng. Eman Al.Swaity Lecture (1): Introduction.
ParCFD Parallel computation of pollutant dispersion in industrial sites Julien Montagnier Marc Buffat David Guibert.
Ch 1. A Python Q&A Session Spring Why do people use Python? Software quality Developer productivity Program portability Support libraries Component.
Center for Component Technology for Terascale Simulation Software CCA is about: Enhancing Programmer Productivity without sacrificing performance. Supporting.
Case Study in Computational Science & Engineering - Lecture 2 1 Parallel Architecture Models Shared Memory –Dual/Quad Pentium, Cray T90, IBM Power3 Node.
Problem Solving with NetSolve Michelle Miller, Keith Moore,
Presented by An Overview of the Common Component Architecture (CCA) The CCA Forum and the Center for Technology for Advanced Scientific Component Software.
Atlas Graphics Group MeetingDec, 1 The Colt Distribution - Open Source Libraries for High Performance Scientific and Technical.
Numerical Libraries Project Microsoft Incubation Group Mary Beth Hribar Microsoft Corporation CSCAPES Workshop June 10, 2008 Copyright Microsoft Corporation,
1CPSD Software Infrastructure for Application Development Laxmikant Kale David Padua Computer Science Department.
ACES WorkshopJun-031 ACcESS Software System & High Level Modelling Languages by
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
1 1 What does Performance Across the Software Stack mean?  High level view: Providing performance for physics simulations meaningful to applications 
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
CCA Common Component Architecture CCA Forum Tutorial Working Group CCA Status and Plans.
Domain Decomposition in High-Level Parallelizaton of PDE codes Xing Cai University of Oslo.
October 2008 Integrated Predictive Simulation System for Earthquake and Tsunami Disaster CREST/Japan Science and Technology Agency (JST)
Cracow Grid Workshop, November 5-6, 2001 Concepts for implementing adaptive finite element codes for grid computing Krzysztof Banaś, Joanna Płażek Cracow.
Connections to Other Packages The Cactus Team Albert Einstein Institute
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA
Automatic Parameterisation of Parallel Linear Algebra Routines Domingo Giménez Javier Cuenca José González University of Murcia SPAIN Algèbre Linéaire.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Advanced User Support for MPCUGLES code at University of Minnesota October 09,
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
Satisfying Requirements BPF for DRA shall address: –DAQ Environment (Eclipse RCP): Gumtree ISEE workbench integration; –Design Composing and Configurability,
Report from LBNL TOPS Meeting TOPS/ – 2Investigators  Staff Members:  Parry Husbands  Sherry Li  Osni Marques  Esmond G. Ng 
Algebraic Solvers in FASTMath Argonne Training Program on Extreme-Scale Computing August 2015.
1 Rocket Science using Charm++ at CSAR Orion Sky Lawlor 2003/10/21.
C OMPUTATIONAL R ESEARCH D IVISION 1 Defining Software Requirements for Scientific Computing Phillip Colella Applied Numerical Algorithms Group Lawrence.
Center for Component Technology for Terascale Simulation Software (CCTTSS) 110 April 2002CCA Forum, Townsend, TN CCA Status, Code Walkthroughs, and Demonstrations.
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA Shirley Moore CPS5401 Fall 2013 svmoore.pbworks.com November 12, 2012.
Adaptive grid refinement. Adaptivity in Diffpack Error estimatorError estimator Adaptive refinementAdaptive refinement A hierarchy of unstructured gridsA.
SimTK 1.0 Workshop Downloads Jack Middleton March 20, 2008.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
VisIt Project Overview
Kai Li, Allen D. Malony, Sameer Shende, Robert Bell
A survey of Exascale Linear Algebra Libraries for Data Assimilation
Programming Models for SimMillennium
GENERAL VIEW OF KRATOS MULTIPHYSICS
Salient application properties Expectations TOPS has of users
Parallel Program Analysis Framework for the DOE ACTS Toolkit
Presentation transcript:

The ACTS Toolkit (What can it do for you?) Osni Marques and Tony Drummond ( LBNL/NERSC )

06/06/2000NUG/ORNL2 What is the ACTS Toolkit? Advanced Computational Testing and Simulation Tools for development of parallel applications 21 tools developed (primarily) at DOE labs originally conceived as autonomous tools ACTS is an “umbrella” project collect tools leverage numerous independently funded projects

06/06/2000NUG/ORNL3 ACTS Project Goals Bring software packages together into a “toolkit” Make the software interoperable Provide consistent application interfaces Promote general solutions to parallel programming needs Promote code reusability Enable large scale applications Impact DOE science

06/06/2000NUG/ORNL4 Recent Successful Cases Scattering in a quantum system of three charged particles (Rescigno, Baertschy, Isaacs and McCurdy, Dec. 24, 1999). Cosmic Microwave Background Analysis, BOOMERanG collaboration, MADCAP code (Apr. 27, 2000).

06/06/2000NUG/ORNL5 NERSC Activities Make ACTS tools available on NERSC platforms Provide technical support Perform independent evaluation of tools Maintain online ACTS information center Identify new users who can benefit from toolkit Work with users to integrate tools into applications

06/06/2000NUG/ORNL6 ACTS Information Online Information center coordinated and integrated with developers resources general information and pointers to documentation User-oriented information Which tool should I use to solve my problem? Which tools are ready for prime time? Reviews: strengths, weaknesses, caveats, etc

06/06/2000NUG/ORNL7 ACTS Support Support at different levels applications code optimization tool selection tool utilization tool installation Leverage with developers Minimize risk to users

06/06/2000NUG/ORNL8 Tools Categorization Numerical  software that implements numerical algorithms Structural (“frameworks”)  software that manages data, communication Infra-structural  runtime, support tools, developer’s bag

06/06/2000NUG/ORNL9 Numerical Tools Aztec: iterative methods for solving sparse linear systems Hypre: collection of advanced preconditioners Opt++: solution of nonlinear optimization problems PETSc: methods for the solution of PDE related problems PVODE: solvers for large systems of ODE’s ScaLAPACK: dense linear algebra computations SuperLU: direct methods for sparse linear systems

06/06/2000NUG/ORNL10 Structural (Frameworks) Global Arrays: portable, distributed array library, shared memory style of programming Overture: library of grid functions which derives from P++ arrays POET (Parallel Object-oriented Environment and Toolkit): allows for “mixing and matching” of components POOMA (Parallel Object-Oriented Methods and Applications): C++ abstraction layer between algorithm and platform (similar to HPF)

06/06/2000NUG/ORNL11 Infra-structural CUMULVS (Collaborative User Migration User Library for Visualization and Steering), PAWS (Parallel Application WorkSpace): computational steering, data post-processing, interactive visualization Globus: infrastructure for high performance distributed computing (computational grids) SILOON (Scripting Interface Languages for Object-Oriented Numerics): scripting features TAU (Tuning and Analysis Utilities): advanced performance analysis and tuning

06/06/2000NUG/ORNL12 Infra-structural (cont.) Tulip: C++ applications with threads, global pointers and other parallel operations ATLAS (Automatically Tuned Linear Algebra Software), PHiPAC (Portable High Performance ANSI C) : automatic generation of optimized numerical software (mainly BLAS) Nexus: multithreading, communication and resource management facilities PADRE (Parallel Asynchronous Data and Routing Engine) : abstracts the details of representing and managing distributed data PETE (Portable Expression Template Engine) : efficient C++ operator overloading through expression templates

06/06/2000NUG/ORNL13 Tool Status at NERSC Installed and supported (modules,T3E and SP) Aztec PETSc ScaLAPACK SuperLU TAU Known to have been used (T3E) CUMULVS Global Arrays Globus POOMA PVODE

06/06/2000NUG/ORNL14 Aztec Solves large sparse linear systems on distributed memory machines Implements Krylov iterative methods (CG, CGS, Bi- CG-Stab, GMRES, TFQMR) Suite of preconditioners (Jacobi, Gauss-Seidel, overlapping domain decomposition with sparse LU, ILU, BILU within domains) Highly efficient, scalable (1000 processors on the “ASCI Red” machine)

06/06/2000NUG/ORNL15 Aztec (applications) TOUGH2 (Transport Of Unsaturated Groundwater and Heat) code, transport simu- lation in porous and fractured media (LBNL). Co-flowing Annular Jet Combuster, a parallel 3D pseudo-transient simulation to steady state operation ; MPSalsa code (SNL).

06/06/2000NUG/ORNL16 Aztec (basic steps) Prepare your linear system distribute the matrix call AZ_transform set up right-hand side and initial guess call AZ_reorder_vec on initial guess and right-hand side selective an iterative solver and a preconditioner call AZ_solve call AZ_invorder_vec on solution

06/06/2000NUG/ORNL17 PETSc Linear and nonlinear solvers Data structures for distributed vectors and sparse matrices 5 matrix storage formats It has been integrated with Aztec 10+ Krylov iterative subspace methods Preconditioners include Jacobi, additive overlapping Schwartz, parallel ILU(0), etc Modular, object design methodology C and Fortran interfaces Can be used as a front end to other ACTS tools

06/06/2000NUG/ORNL18 PETSc (applications) Multiphase flow, 4 million cell blocks, 32 million DOF, over 10.6 Gflops on an IBM SP (128 nodes), entire simulation runs in less than 30 minutes (Pope, Gropp, Morgan, Seperhrnoori, Smith and Wheeler). Prometheus code (unstructured meshes in solid mechanics), 26 million DOF, 640 nodes on NERSC’s Cray T3E (Adams and Demmel).

06/06/2000NUG/ORNL19 ScalaPACK Parallel version of LAPACK Based on BLACS (Basic Linear Algebra Communication Subroutines) Dense linear algebra (BLAS) Direct solution of linear systems Dense matrix eigensolvers

06/06/2000NUG/ORNL20 ScaLAPACK (applications) Induced current (white arrows) and charge density (colored plane and gray surface) in crystallized glycine due to an external field (Louie, Yoon, Pfrommer and Canning). Cosmic Microwave Background Analysis, BOOMERanG collaboration, MADCAP code (Apr. 27, 2000).

06/06/2000NUG/ORNL21 SuperLU Direct solution of large sparse linear systems Shared and distributed memory implementations Attained 8.3 Gflops on 512 nodes of the T3E

06/06/2000NUG/ORNL22 SuperLU (applications) Scattering in a quantum system of three charged particles (Rescigno, Baertschy, Isaacs and McCurdy, Dec. 24, 1999). SuperLU speedup (matrices dimension varying from to ).

06/06/2000NUG/ORNL23 TAU Profiling of Fortran 90, C, C++, HPF, and HPC++ codes Detailed information (much more than prof/gprof) C++: per-class and per-instance profiling Graphical display of profiling results (built-in viewers, interface to Vampir)

06/06/2000NUG/ORNL24 TAU (Main Control Window) COSY: COmpile manager Status displaY FANCY: File ANd Class displaY CAGEY: CAll Graph Extended displaY CLASSY: CLASS hierarchY browser RACY: Routine and data ACcess profile displaY SPEEDY: Speedup and Parallel Execution Extrapolation DisplaY

06/06/2000NUG/ORNL25 TAU (SPEEDY)

06/06/2000NUG/ORNL26 Coming up Hypre PVODE POOMA Globus PETE (summer student) Survey of iterative linear solver packages

06/06/2000NUG/ORNL27 PVODE and POOMA PVODE is being used in the ParFlow ground- water flow project to solve three-dimensional models of fluid flow and chemical transport through heterogeneous porous media (LNNL). MC++ Monte Carlo neutron transport simulation written using the Tecolote framework, which is built atop POOMA, (ASCI Blanca project, LANL).

06/06/2000NUG/ORNL28 Globus Provides means for creating a large scale computational grid A toolkit of core services with which applications can be developed to access the grid. Operates with other ACTS tools (like CUMULVS, PAWS and Nexus)

06/06/2000NUG/ORNL29 Globus (cont.) NASA’s Information Power Grid (IPG) joins supercomputers and storage devices owned by participating organizations into a single, seamless computing environment. With a Globus-enabled X-ray microto- mography program a fully 3-D reconstruction of an ant head (2mm in diameter) was obtained in less than 10 minutes using the acquisition hardware at a beamline at ANL and an SGI Origin 2000 at NPACI.

06/06/2000NUG/ORNL30 Future Directions CCA (Common Component Architecture) Developing standardized ways of managing numerical components to allow mixing-and-matching Frameworks for gluing components together Similarities to CORBA, DCOM, Java Beans Scientific interface description language (allowing Fortran) ESI (Equation Solver Interface) Developing standardized interfaces for scalable linear solvers Specific test case for CCA component design

acts-