Welcome to Indiana University Clusters

Slides:



Advertisements
Similar presentations
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Advertisements

Introduction to HPC Workshop October Introduction Rob Lane HPC Support Research Computing Services CUIT.
High Performance Computing
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
Introduction to RCC for Intro to MRI 2014 July 25, 2014.
ITCS 6/8010 CUDA Programming, UNC-Charlotte, B. Wilkinson, Jan 22, 2011assignprelim.1 Assignment Preliminaries ITCS 6010/8010 Spring 2011.
Introduction to UNIX/Linux Exercises Dan Stanzione.
ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson, 2012, Jan 18, 2012assignprelim.1 Assignment Preliminaries ITCS 4145/5145 Spring 2012.
Systems Software & Operating systems
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
How to get started on cees Mandy SEP Style. Resources Cees-clusters SEP-reserved disk20TB SEP reserved node35 (currently 25) Default max node149 (8 cores.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
Running Parallel Jobs Cray XE6 Workshop February 7, 2011 David Turner NERSC User Services Group.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Introduction to Parallel Computing Presented by The Division of Information Technology Computer Support Services Department Research Support Group.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
Setting up visualization. Make output folder for visualization files Log into vieques $ ssh
Assignprelim.1 Assignment Preliminaries © 2012 B. Wilkinson/Clayton Ferner. Modification date: Jan 16a, 2014.
Using ROSSMANN to Run GOSET Studies Omar Laldin ( using materials from Jonathan Crider, Harish Suryanarayana ) Feb. 3, 2014.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
Grid Computing: An Overview and Tutorial Kenny Daily BIT Presentation 22/09/2016.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
IPPP Grid Cluster Phil Roffe David Ambrose-Griffith.
Editing, Transferring, and Running Files on Vieques Daniel Malmer Dowell Lab Short Reads Course 6/9/15.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Workstations & Thin Clients
bitcurator-access-webtools Quick Start Guide
GRID COMPUTING.
Specialized Computing Cluster An Introduction
Auburn University
Development Environment
Unix Scripts and PBS on BioU
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
HPC usage and software packages
Welcome to Indiana University Clusters
UNDERSTANDING YOUR COMPUTER
How to use the HPCC to do stuff
Chapter 5: Using System Software
Using Paraguin to Create Parallel Programs
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
CommLab PC Cluster (Ubuntu OS version)
BIMSB Bioinformatics Coordination
Assignment Preliminaries
Practice #0: Introduction
Postdoctoral researcher Department of Environmental Sciences, LSU
Intro to UNIX System and Homework 1
Welcome to our Nuclear Physics Computing System
Introduction to Linux and Supercomputers
Paul Sexton CS 566 February 6, 2006
Introduction to HPC Workshop
College of Engineering
Compiling and Job Submission
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Welcome to our Nuclear Physics Computing System
High Performance Computing in Bioinformatics
Michael P. McCumber Task Force Meeting April 3, 2006
Queueing System Peter Wad Sackett.
Introduction to High Performance Computing Using Sapelo2 at GACRC
Software - Operating Systems
Quick Tutorial on MPICH for NIC-Cluster
Starter Which of these inventions is: Used most by people in Britain
Working in The IITJ HPC System
Windows 10 An Operating System
Presentation transcript:

Welcome to Indiana University Clusters Cluster Quick Guide Welcome to Indiana University Clusters

IU Compute Clusters Computers you can log into karst.uits.iu.edu 256 nodes 16 cores 16-32GB ram per core for general use IU Compute Clusters How to log in: Mac – open terminal: ssh username@computer i.e.: ssh hpstrn29@karst.uits.iu.edu Windows… get putty! mason.indiana.edu 18 nodes 32 cores per node 512GB ram per node for high memory genomics jobs. Computers you can log into karst.uits.iu.edu 256 nodes, 16 cores, 16GB ram per core, for general use mason.indiana.edu 18 nodes, 32 cores per node, 512GB ram per node, for high memory genomics jobs. bigred2.uits.iu.edu (slightly different architecture) ~23,000 cores, ~43,000 GB ram total, for really big jobs

hpstrn29@karst.uits.iu.edu

How much space do I have? You can check it with: [hpstrn21@h2 ~]$ quota Disk quotas for user hpstrn29 Filesystem usage quota % computed Home 2G 100G 2.0% Now dc2/scratch 0.00G none -- 2017-05-31 Sda 0G 50.00T 0.0% 2017-07-28 You can get to your scratch space with cd $SCRATCH – this is open between accounts! Shared folder for the course: cd $MDIBL

Software Modules Modules make certain programs available to you in your path without having to find/install them. View module list (or see http://ncgas.org/software-list.php): module av List the modules you have loaded: module list Determine if a module needs another module loaded first (dependency) and to get more information on the module in general: module display <module name> Load a module: module load <module name> Unload a module (if you for some reason need to): module unload <module name>

Other Software BUT THIS ONLY WORKS FOR SOURCE CODE!!!! PYTHON AND R ARE STORED ELSEWHERE! What R libraries are installed? module load r R installed.packages() What Python modules are installed? module load python #should be loaded for you pip list SEE ME if you need/want to add other python/R libraries!

Job Submission You can also open an interactive que with: In order to run a job for more than 20 minutes, you need to submit the job to the que. In order to submit to the que, you need to make a job file. #PBS -k oe #keep output and error #PBS -m abe #mail me when job : a – abort, b - begins, e - ends #PBS -M <your email> #PBS -N <name of job> #PBS -l nodes=1:ppn=1,vmem=16gb,walltime=2:00:00 #Load modules required module load java/1.7.0_40 module load fastqc #you must enter the directory where your data is, with an absolute path cd /N/dc2/scratch/ss93/Analysis #call the program as you would normally fastqc NewData.fq You can also open an interactive que with: qsub –I –l nodes=1:ppn=1,vmem=1gb,walltime=10:00:00 This allows you to interact with the compute node (not log in)!

Quick tips How much resource do I request? No software you are using can communicate between nodes – so only request ONE. Unless you can specify how many CPUs to use for a software, it won’t run across different processors – request ONE. The more processors you request… the longer you wait to run it (not so much an issue here if people share nicely) There are only 16GB of memory per NODE. If you request more, it won’t run. If you request 16GB, it can’t run basic things that a computer needs to do (OS!) and will often die. HOW DO I KNOW? Read the email that you get back!

Quick tips HOW DO I KNOW? Read the email that you get back!

Managing your jobs What is the status of my job (for example if you are me/ss93), or what is the job id? qstat -u ss93 NOTE: q - waiting in the que, r – running, c - complete Delete a job (the number is listed when you use qstat): qdel -j 60123 When will my job run? showstart 60123 What's going on with my job? Your job will create two files, based on the name you provided in the –N flag. NAME.o<jobnumber> - output that would normally print to STDOUT (screen) NAME.e<jobnumber> - output that would normally print to STDERR (usually screen)

Other tips Try running commands before putting them in a job file. This ensure they will run correctly. You can always stop them with ctrl c. 
 Create a blank job file and just copy it to another file name when you are creating a new job submission. Have some sort of naming paradigm for jobs. I call all mine RunWhatever. Some people give them a .job extension. It doesn't matter what you do, but it helps when you are looking for them in your folders. You can run multiple things in parallel if you put an ‘&’ at the end of each command and then ‘wait’ at the end of the file. Don’t forget the ’wait’!! For more in depth information, try the KnowledgeBase: https://kb.iu.edu/index.html