Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013.

Slides:



Advertisements
Similar presentations
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Job Submission.
Advertisements

Chapter 3. MPI MPI = Message Passing Interface Specification of message passing libraries for developers and users –Not a library by itself, but specifies.
Using the Argo Cluster Paul Sexton CS 566 February 6, 2006.
Job Submission Using PBSPro and Globus Job Commands.
Koç University High Performance Computing Labs Hattusas & Gordion.
HCC Workshop Department of Earth and Atmospheric Sciences September 23/30, 2014.
Network for Computational Nanotechnology (NCN) Purdue, Norfolk State, Northwestern, UC Berkeley, Univ. of Illinois, UTEP Basic Portable Batch System (PBS)
HPCC Mid-Morning Break MPI on HPCC Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Research
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
PBS Job Management and Taskfarming Joachim Wagner
Batch Queuing Systems The Portable Batch System (PBS) and the Load Sharing Facility (LSF) queuing systems share much common functionality in running batch.
Using Clusters -User Perspective. Pre-cluster scenario So many different computers: prithvi, apah, tejas, vayu, akash, agni, aatish, falaq, narad, qasid.
Running Jobs on Jacquard An overview of interactive and batch computing, with comparsions to Seaborg David Turner NUG Meeting 3 Oct 2005.
IT MANAGEMENT OF FME, 21 ST JULY  THE HPC FACILITY  USING PUTTY AND WINSCP TO ACCESS THE SERVER  SENDING FILES TO THE SERVER  RUNNING JOBS 
High Performance Computing
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh ssh.fsl.byu.edu You will be logged in to an interactive node.
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
ISG We build general capability Purpose After this tutorial, you should: Be comfortable submitting work to the batch queuing system of olympus and be familiar.
JGI/NERSC New Hardware Training Kirsten Fagnan, Seung-Jin Sul January 10, 2013.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
Blaise Barney, LLNL ASC Tri-Lab Code Development Tools Workshop Thursday, July 29, 2010 Lawrence Livermore National Laboratory, P. O. Box 808, Livermore,
Electronic Visualization Laboratory, University of Illinois at Chicago MPI on Argo-new Venkatram Vishwanath Electronic Visualization.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
Rsv-control Marco Mambelli – Site Coordination meeting October 1, 2009.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Swarm on the Biowulf2 Cluster Dr. David Hoover, SCB, CIT, NIH September 24, 2015.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Network Queuing System (NQS). Controls batch queues Only on Cray SV1 Presently 8 queues available for general use and one queue for the Cray analyst.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
© 2014 IBM Corporation SLURM for Yorktown Bluegene/Q.
APST Internals Sathish Vadhiyar. apstd daemon should be started on the local resource Opens a port to listen for apst client requests Runs on the host.
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat.
Software Tools Using PBS. Software tools Portland compilers pgf77 pgf90 pghpf pgcc pgCC Portland debugger GNU compilers g77 gcc Intel ifort icc.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
1 HPCI Presentation Kulathep Charoenpornwattana. March 12, Outline Parallel programming with MPI Running MPI applications on Azul & Itanium Running.
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Event Service Wen Guan University of Wisconsin 1.
How to use HybriLIT Matveev M. A., Zuev M.I. Heterogeneous Computations team HybriLIT Laboratory of Information Technologies (LIT), Joint Institute for.
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
Modules, Compiling WRF, and Running on CHPC Clusters Adam Varble WRF Users Meeting 10/26/15.
Debugging Lab Antonio Gómez-Iglesias Texas Advanced Computing Center.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarksEGEE-III INFSO-RI MPI on the grid:
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
Outline Introduction/Questions
Hands on training session for core skills
Welcome to Indiana University Clusters
PARADOX Cluster job management
Unix Scripts and PBS on BioU
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
Welcome to Indiana University Clusters
How to use the HPCC to do stuff
Using Paraguin to Create Parallel Programs
Joker: Getting the most out of the slurm scheduler
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
Short Read Sequencing Analysis Workshop
Paul Sexton CS 566 February 6, 2006
College of Engineering
Compiling and Job Submission
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
High Performance Computing in Bioinformatics
Quick Tutorial on MPICH for NIC-Cluster
Working in The IITJ HPC System
Short Read Sequencing Analysis Workshop
Presentation transcript:

Introduction to Using SLURM on Discover Chongxun (Doris) Pan September 24, 2013

SLURM ON! Why Simple Linux Utility for Resource Management? An open source, fault-tolerant, and highly scalable cluster workload scheduler What to expect after we migrate from PBS to SLURM next month? Nearly Transparent. Most of your PBS scripts should run on SLURM without any changes. A few things to bear in mind... What if you cannot wait to go native SLURM? Kudos for the forward-thinking attitude! Will show you how easy it is to convert to native SLURM 2

What to Expect NCCS is working HARD to develop wrappers to help make conversion transparent. But each workload scheduler has its unique features… Most of the PBS commands, including qsub, qstat, qdel, qalter, should function the same way as in PBS, along with their typical options “qalter –o” is not supported with the current release. With SLURM, as soon as a job starts the output is written directly to the file that you specified in the “-o” directive. No temporary pbs_spool files! You may notice slight formatting difference for “qstat” output. Interactive batch using “qsub –I” works the same way, so does “xsub –I”. 3

What to Expect (Cont’d) For batch jobs, most of the commonly used PBS environment variables are defined, including PBS_NODEFILE, PBS_JOBID, PBS_JOBDIR, and PBS_O_WORKDIR The PBS job scripts will run almost seamlessly in SLURM. Most of the PBS attributes work the same way, including #PBS –l select=i:ncpus=j:mpiprocs=k:procs=west/sand,walltime=hh:mm:s s #PBS -W group_list, -W depend, -W umask #PBS –o, –e #PBS –j is ignored. By default, SLURM writes stdout and stderr to one file, as specified by #PBS -o. To separate stdout and stderr, specify both #PBS –o and -e. If only #PBS -e is specified, the stdout will be written into a file named, slurm-$SLURM_JOBID.out. Without –e and –o, both the stdout and stderr will be written into slurm-$SLURM_JOBID.out. 4

What to Expect (Cont’d) For Intel MPI, the following Intel MPI modules are available and “mpirun” command works the same way as under PBS: mpi/impi mpi/impi mpi/impi mpi/impi For MVAPICH2 and OpenMPI, “mpirun” should work seamlessly as well in the SLURM environment. We intend to retire some older and all the beta versions of Intel MPI. Please contact NCCS User Support for assistance in migrating to a newer version of We highly recommend switching to Intel MPI 4 modules and using “mpiexec.hydra” instead of “mpirun” for better performance. 5

SLURM and Older Intel MPI, MVAPICH, and OpenMPI Versions Intel MPI mpi/impi-3.1.* mpi/impi mpi/impi mpi/impi-4.*-beta MVAPICH other/mpi/mvapich2- 1.6rc2/intel other/mpi/mvapich2-1.7* other/mpi/mvapich2-1.8/* other/mpi/mvapich2-1.8a2/* OpenMPI mpi/openmpi-1.2.5/intel-9 mpi/openmpi-1.2.5/intel-10 other/mpi/openmpi/1.4.* other/mpi/openmpi/1.6-gcc_ _nag other/mpi/openmpi/1.6.0* other/mpi/openmpi/1.6.3* other/mpi/openmpi/1.6.4* 6 These old (and beta) versions are incompatible with SLURM. If you use the versions below, please contact ASAP for help in migrating to SLURM-compatible

Native SLURM Basics Six basic user commands should cover most of your needs sbatch -- Submit job script (Batch mode) salloc -- Create job allocation and start a shell (Interactive Batch mode) srun – Run a command within a batch allocation that was created by sbatch or salloc scancel -- Delete jobs from the queue squeue -- View the status of jobs sinfo – View information on nodes and queues --help option prints brief description of all options --usage option prints a list of the options “srun” or mpirun/mpiexec.hydra? srun is the best integrated with SLURM and supports process tracking, accounting, task affinity, suspend/resume and other features. 7

SLURM commands vs. PBS commands 8 sbatch myscript.jqsub myscript.j salloc -N 2 -n 16 –ntasks-per- node=8 -t 1:00:00 -p general --account=k3001 qsub –I -l select=2:mpiprocs=8,walltime=1:00:00 –q general -W group_list=k3001 squeueqstat squeue –j qstat –f sinfoqstat -Q squeue –u cpan2qstat –u cpan2 scancel qdel For sbatch or salloc commands, almost all options have two formats: ● A single letter option (e.g. “-p debug”, “-N 2”, “-n 16”), or ● A verbose option (e.g. “--partition=debug”, “--nodes=2”, “--ntasks=16”) Acceptable walltime “-t or –time=“formats include "minutes", "minutes:seconds", "hours:minutes:seconds”.

SLURM environment variables vs. PBS environment variables 9 SLURM_JOBIDPBS_JOBID SLURM_NODELISTPBS_NODEFILE SLURM_SUBMIT_HOSTPBS_O_HOST SLURM_SUBMIT_DIRPBS_O_WORKDIR SLURM_NNODES (number of nodes requested)N/A SLURM_NTASKS (number of MPI procs requested) N/A SLURM_CPUS_ON_NODE (number of CPUS on the allocated node) N/A SLURM_NTASKS_PER_NODE (return “mpiprocs” number) N/A

A native SLURM job script vs. A PBS job script 10 #!/bin/csh #SBATCH -J MyFirstSLURM #SBATCH -A k3001 #SBATCH -N 2 -n 16 –ntasks-per-node=8 #SBATCH –t 1:00:00 #SBATCH #SBATCH -p general #SBATCH -o testmpi.out ##SBATCH -e testmpi.err #SBATCH –d afterany:11697 module purge module load comp/intel module load mpi/impi cd $NOBACKUP mpif90 testmpi.f90 mpiexec.hydra -np 16./a.out #!/bin/csh #PBS -N MyFirstPBS #PBS -W group_list=k3001 #PBS -l select=2:mpiprocs=8 #PBS -l walltime=1:10:00 #PBS -M #PBS -q general #PBS -o testmpi.output #PBS –j oe #PBS -W depend=afterany:11697 module purge module load comp/intel module load mpi/impi cd $NOBACKUP mpif90 testmpi.f90 mpiexec.hydra -np 16./a.out

To Learn More About SLURM… slurm.schedmd.com/ 11