Welcome to Indiana University Clusters

Slides:



Advertisements
Similar presentations
GXP in nutshell You can send jobs (Unix shell command line) to many machines, very fast Very small prerequisites –Each node has python (ver or later)
Advertisements

Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
Introduction to HPC Workshop October Introduction Rob Lane HPC Support Research Computing Services CUIT.
High Performance Computing
OpenVMS System Management A different perspective by Andy Park TrueBit b.v.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Quick Tutorial on MPICH for NIC-Cluster CS 387 Class Notes.
Introduction to RCC for Intro to MRI 2014 July 25, 2014.
HPCC Mid-Morning Break Powertools Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Research
Introduction to UNIX/Linux Exercises Dan Stanzione.
Thrive Installation.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
Electronic Visualization Laboratory, University of Illinois at Chicago MPI on Argo-new Venkatram Vishwanath Electronic Visualization.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
TATA Motors Limited Commercial Vehicles Business Unit TML CVBU eLearning Application Installation Guide for application installer and update patches.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
How to get started on cees Mandy SEP Style. Resources Cees-clusters SEP-reserved disk20TB SEP reserved node35 (currently 25) Default max node149 (8 cores.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
Advanced topics Cluster Training Center for Simulation and Modeling September 4, 2015.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
NREL is a national laboratory of the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, operated by the Alliance for Sustainable.
Setting up visualization. Make output folder for visualization files Log into vieques $ ssh
Using ROSSMANN to Run GOSET Studies Omar Laldin ( using materials from Jonathan Crider, Harish Suryanarayana ) Feb. 3, 2014.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
Grid Computing: An Overview and Tutorial Kenny Daily BIT Presentation 22/09/2016.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
IPPP Grid Cluster Phil Roffe David Ambrose-Griffith.
Editing, Transferring, and Running Files on Vieques Daniel Malmer Dowell Lab Short Reads Course 6/9/15.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Workstations & Thin Clients
GRID COMPUTING.
Specialized Computing Cluster An Introduction
Auburn University
Welcome to Indiana University Clusters
Development Environment
Unix Scripts and PBS on BioU
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
HPC usage and software packages
Discussion #11 11/21/16.
How to use the HPCC to do stuff
Discussion 11 Final Project / Git.
IW2D migration to HTCondor
Using Paraguin to Create Parallel Programs
ASU Saguaro 09/16/2016 Jung Hyun Kim.
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
CommLab PC Cluster (Ubuntu OS version)
BIMSB Bioinformatics Coordination
Introduction to Programming the WWW I
Assignment Preliminaries
Practice #0: Introduction
Lab 1 introduction, debrief
Intro to UNIX System and Homework 1
Short Read Sequencing Analysis Workshop
Welcome to our Nuclear Physics Computing System
Paul Sexton CS 566 February 6, 2006
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Welcome to our Nuclear Physics Computing System
High Performance Computing in Bioinformatics
Michael P. McCumber Task Force Meeting April 3, 2006
Queueing System Peter Wad Sackett.
Introduction to High Performance Computing Using Sapelo2 at GACRC
Quick Tutorial on MPICH for NIC-Cluster
Starter Which of these inventions is: Used most by people in Britain
Short Read Sequencing Analysis Workshop
Presentation transcript:

Welcome to Indiana University Clusters Cluster Quick Guide Welcome to Indiana University Clusters

Computers you can log into karst.uits.iu.edu – main use 256 nodes, 16 cores, 16GB ram per core, for general jobs we have reserved time on this machine Wednesday through Friday! 3 nodes = 48cores, 48GB ram mason.indiana.edu 18 nodes, 32 cores per node, 512GB ram per node, for high memory genomics jobs. This is for high memory jobs only – likely only for STAR. We do not have reserved time on this machine, and you will have to wait in que. Log In: ssh username@computer i.e.: ssh ss93@karst.uits.iu.edu

How much space do I have? You can check it with: [hpstrn21@h2 ~]$ quota Disk quotas for user hpstrn21 Filesystem usage quota % computed Home 2G 100G 2.0% Now dc2/scratch 0.00G none -- 2016-05-31 Sda 0G 50.00T 0.0% 2016-07-28

Modules Clusters are shared resources – Mason/Karst/BigRed2 are used world wide! Problem: Everyone keeps asking for different versions of software for different programs that depend on them… Solution: Modules! Modules allow you to easily add software to your environment depending on your needs. You can swap between different versions of R, python, etc. and whatever you have loaded does not affect anyone else. For example: I want R 3.1.1 for this class! module load r (3.1.1 is the default) #all modules on our system are lowercase. This is semi-standard But I need to use R 3.0.1 for this other course that requires it! module unload r (different versions cannot be installed at the same time!) module load r/3.0.1

Modules Modules make certain programs available to you in your path without having to find/install them. View module list: module av List the modules you have loaded: module list Determine if a module needs another module loaded first (dependency) and to get more information on the module in general: module display <module name> Load a module: module load <module name> Unload a module (if you for some reason need to): module unload <module name> If there are modules you will be using every time you log in, add them to your ~/.module file! i.e: add “module load r “

Job Submission You can also open an interactive que with: In order to run a job for more than 20 minutes, you need to submit the job to the que. In order to submit to the que, you need to make a job file. #PBS -k oe #keep output and error #PBS -m abe #mail me when job : a – abort, b - begins, e - ends #PBS -M <your email> #PBS -N <name of job> #PBS -l nodes=1:ppn=1,vmem=16gb,walltime=2:00:00 #Load modules required module load java/1.7.0_40 module load fastqc #you must enter the directory where your data is, with an absolute path cd /N/dc2/scratch/ss93/Analysis #call the program as you would normally fastqc NewData.fq You can also open an interactive que with: qsub –I –l nodes=1:ppn=1,vmem=16gb,walltime=10:00:00 This allows you to interact with the compute node (not log in)!

Managing your jobs What is the status of my job (for example if you are me/ss93), or what is the job id? qstat -u ss93 NOTE: q - waiting in the que, r – running, c - complete Delete a job (the number is listed when you use qstat): qdel -j 60123 What's going on with my job? Your job will create two files, based on the name you provided in the –N flag. NAME.o<jobnumber> - output that would normally print to STDOUT (screen) NAME.e<jobnumber> - output that would normally print to STDERR (usually screen)

Tips Try running commands before putting them in a job file. This ensure they will run correctly. You can always stop them with ctrl c. 
 Create a blank job file and just copy it to another file name when you are creating a new job submission. Have some sort of naming paradigm for jobs. I call all mine RunWhatever. Some people give them a .job extension. It doesn't matter what you do, but it helps when you are looking for them in your folders. You can run multiple things in parallel if you put an ‘&’ at the end of each command and then ‘wait’ at the end of the file. Don’t forget the ’wait’!! For more in depth information, try the KnowledgeBase: https://kb.iu.edu/index.html

SIGN INs: Access to: Rstudio.iu.edu – rstudio on cluster Galaxy.iu.edu – NCGAS galaxy installation (runs on Mason!)