Critical Flags, Variables, and Other Important ALCF Minutiae Jini Ramprakash Technical Support Specialist Argonne Leadership Computing Facility.

Slides:



Advertisements
Similar presentations
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Job Submission.
Advertisements

Network for Computational Nanotechnology (NCN) Purdue, Norfolk State, Northwestern, UC Berkeley, Univ. of Illinois, UTEP Advanced Portable Batch System.
Using the Argo Cluster Paul Sexton CS 566 February 6, 2006.
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
PBS Job Management and Taskfarming Joachim Wagner
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
High Performance Computing
Job Submission on WestGrid Feb on Access Grid.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh ssh.fsl.byu.edu You will be logged in to an interactive node.
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
ORNL is managed by UT-Battelle for the US Department of Energy Tools Available for Transferring Large Data Sets Over the WAN Suzanne Parete-Koon Chris.
Plans for Exploitation of the ORNL Titan Machine Richard P. Mount ATLAS Distributed Computing Technical Interchange Meeting May 17, 2013.
ISG We build general capability Purpose After this tutorial, you should: Be comfortable submitting work to the batch queuing system of olympus and be familiar.
MIGRATING TO THE SHARED COMPUTING CLUSTER (SCC) SCV Staff Boston University Scientific Computing and Visualization.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
AQS Web Quick Reference Guide Changing Raw Data Values Using Maintenance 1. From Main Menu, click Maintenance, Sample Values, Raw Data 2. Enter monitor.
Resource Management and Accounting Working Group Working Group Scope and Components Progress made Current issues being worked Next steps Discussions involving.
Sun Grid Engine. Grids Grids are collections of resources made available to customers. Compute grids make cycles available to customers from an access.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Using The Cluster. What We’ll Be Doing Add users Run Linpack Compile code Compute Node Management.
Copyright © 2007, Oracle. All rights reserved. Managing Concurrent Requests.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
How to get started on cees Mandy SEP Style. Resources Cees-clusters SEP-reserved disk20TB SEP reserved node35 (currently 25) Default max node149 (8 cores.
Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013.
Network Queuing System (NQS). Controls batch queues Only on Cray SV1 Presently 8 queues available for general use and one queue for the Cray analyst.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
Argonne Leadership Computing Facility ALCF at Argonne  Opened in 2006  Operated by the Department of Energy’s Office of Science  Located at Argonne.
Data Transfers in the ALCF Robert Scott Technical Support Analyst Argonne Leadership Computing Facility.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
Threaded Programming Lecture 2: Introduction to OpenMP.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
CERN Running a LCG-2 Site – Oxford July - 1 LCG2 Administrator’s Course Oxford University, 19 th – 21 st July Developed.
Introduction to Hartree Centre Resources: IBM iDataPlex Cluster and Training Workstations Rob Allan Scientific Computing Department STFC Daresbury Laboratory.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Cliff Addison University of Liverpool NW-GRID Training Event 26 th January 2007 SCore MPI Taking full advantage of GigE.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
Using ROSSMANN to Run GOSET Studies Omar Laldin ( using materials from Jonathan Crider, Harish Suryanarayana ) Feb. 3, 2014.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Information Technology Services Brett D. Estrade, LSU – High Performance Computing Phone:
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Specialized Computing Cluster An Introduction
Welcome to Indiana University Clusters
PARADOX Cluster job management
HPC usage and software packages
Welcome to Indiana University Clusters
How to use the HPCC to do stuff
Chapter 2: System Structures
Architecture & System Overview
CommLab PC Cluster (Ubuntu OS version)
Short Read Sequencing Analysis Workshop
Paul Sexton CS 566 February 6, 2006
Compiling and Job Submission
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Requesting Resources on an HPC Facility
Sun Grid Engine.
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Short Read Sequencing Analysis Workshop
Presentation transcript:

Critical Flags, Variables, and Other Important ALCF Minutiae Jini Ramprakash Technical Support Specialist Argonne Leadership Computing Facility

Presentation outline  It’s all about your job! –Job management –Job basics Submission Queuing Execution Termination  Software environment  Optimization for beginners  ALCF resources, outlined Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 2

Job management  Cobalt (the ALCF resource scheduler) is used on all ALCF systems –Similar to PBS but not the same –Find more information at  Job management commands: –qsub: submit a job –qstat: query a job status –qdel: delete a job –qalter: alter batched job parameters –qmove: move job to different queue –qhold: place queued (non-running) job on hold –qrls: release hold on job –showres: show current and future reservations Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 3

Job basics – submission  Two modes of submitting jobs –Basic –Script mode  Get all flags and options by running ‘man qsub’  For example:  qsub -A alchemy -n mode c1 -t env “OMP_NUM_THREADS=4” lead_to_gold –In English: Charge project “Alchemy” for this job. Run on nodes, with one MPI rank per node. Run for 720 minutes. Set the “OMP_NUM_THREADS” environment variable to 4. Run the “lead_to_gold” binary. Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 4

qsub checks your submission for sanity  Did you specify a nodecount and walltime? Are they legal?  Is the mode you specified valid?  Did you ask for more than the minimum runtime?  Are you a member of the project you specified? Does that project have a usable allocation?  If so … all systems go! Get a JOBID, and put it in the queue Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 5

Not there yet! Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 6

Job basics - life in the queue  Periodically, your job’s score will increase  Periodically, the scheduler will decide if there are any jobs it wants to run  Check current state with qstat  At some point, your score will be high enough, and it will be YOUR TURN! Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 7

Score accrual  Large jobs are prioritized  Jobs that have been waiting long are prioritized  INCITE/ALCC projects are prioritized  Negative allocations have a score cap lower than the starting score of other jobs Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 8

Job basics - execution  Book-keeping –Put a start record in the database. Output a log file start record. Send of job start if –notify was requested. Start job timers  Fire up to execute the job –Cobalt boots partition –runjob starts executable Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 9

Script mode jobs  All jobs launch via runjob on the service nodes  Script mode jobs launch your script on a special login node  That script is responsible for calling runjob to launch the actual compute-node job  You are charged for the duration of the script Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 10

Job basics – termination aka are we there yet?  Your requested wall-time ticks down. Either your runjob returns, or you run out of wall-time and your job is forcibly removed  Job-end cleanup happens –If your partition wasn’t cleaned up, that happens now  Job-end book-keeping happens –Database, log file, notify if requested Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 11

Job basics – Termination, life after your job  If you had a job depending on you, it can be released to run. If you had a non-zero exit code, it moves to dep_fail instead  That night, the log files will be fed into clusterbank (the ALCF accounting system) to create charges Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 12

Non-standard job events  Reservations and/or draining  qsub rejection  Job holds  Job redefinition (qalter)  Job removal (qdel)  Abnormal job failure  Why isn’t this job running? Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 13

Software environment - SoftEnv  A tool for managing your environment –Sets your PATH to access desired front-end tools –Your compiler version can be changed here  Settings: –Maintained in the file ~/.soft –Add/remove keywords from ~/.soft to change environment –Make is at the very end  Commands: –softenv a list of all keywords defined on the systems –resoft reloads initial environment from ~/.soft file –soft add|remove keyword Temporarily modify environment by adding/removing keywords  Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 14

Software libraries  ALCF Supports two sets of libraries: –IBM system and provided libraries: /bgsys/drivers/ppcfloor glibc mpi –Site supported libraries and programs: /soft/ PETSc ESSL –And many others See Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 15

Compiler wrappers Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 16  MPI wrappers for IBM XL cross-compilers:  MPI wrappers for GNU cross-compilers: WrapperThread-Safe Wrapper Underlying Compiler Description mpixlcmpixlc_rbgxlcIBM BG C Compiler mpixlcxxmpixlcxx_rbgxlCIBM BG C++ Compiler mpixlf77mpixlf77_rbgxlfIBM BG Fortran 77 Compiler mpixlf90mpixlf90_rbgxlf90IBM BG Fortran 90 Compiler mpixlf95mpixlf95_rbgxlf95IBM BG Fortran 95 Compiler mpixlf2003mpixlf2003_rbgxlf2003IBM BG Fortran 2003 Compiler WrapperUnderlying CompilerDescription mpiccpowerpc-bgp-linux-gccGNU BG C Compiler mpicxxpowerpc-bgp-linux-g++GNU BG C++ Compiler mpif77powerpc-bgp-linux-gfortranGNU BG Fortran 77 Compiler mpif90powerpc-bgp-linux-gfortranGNU BG Fortran 90 Compiler

Optimization for beginners  Suggested set of optimization levels from least to most optimization:  -O0 # best level for use with a debugger  -O2 # good level for verifying correctness, baseline perf  -O2 -qmaxmem=-1 -qhot=level=0  -O3 -qstrict (preserves program semantics)  -O3  -O3 -qhot=level=1  -O4  -O5 Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 17

Optimization tips  -qlistopt generates a listing with all flags used in compilation  -qreport produces a listing, shows how code was optimized  Performance can decrease at higher levels of optimization, especially at -O4 or -O5  May specify different optimization levels for different routines/files Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 18

ALCF Resources – BG/Q systems  Mira – BG/Q system –49,152 nodes / 786,432 cores –786 TB of memory –Peak flop rate: 10 PF –Linpack flop rate: 8.1 PF  Cetus (T&D) – BG/Q system –1024 nodes / 16,384 cores –16 TB of memory –Peak flop rate: 208 TF  Vesta (T&D) -­‐ BG/Q systems –2,048 nodes / 32,768 cores –32 TB of memory –Peak flop rate: 416 TF Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 19

ALCF Resources – supporting systems  Tukey –Nvidia system –100 nodes / 1600 x86 cores/ 200 M2070 GPUs –6.4 TB x86 memory / 1.2 TB GPU memory –Peak flop rate: 220 TF  Storage –Scratch: 28.8 PB raw capacity, 240 GB/s bw (GPFS) –Home: 1.8 PB raw capacity, 45 GB/s bw (GPFS) –Storage upgrade planned in 2015 Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 20

ALCF Resources Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 21 Mira 48 racks/768K cores 10 PF Cetus (Dev) 1 rack/16K cores 208 TF Tukey (Viz) 100 nodes/1600 cores 200 NVIDIA GPUs 220 TF Networks 100Gb (via Esnet, internet2 UltraScienceNet) Vesta (Dev) 2 racks/32K cores 416 TF

Coming up next…  Data Transfers in the ALCF - Robert Scott, ALCF Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 22

Thank You!  Questions? Argonne Leadership Computing Facility - supported by the Office of Science of the U.S. Department of Energy 23