ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.

Slides:



Advertisements
Similar presentations
Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Advertisements

including File Management
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Blackbird: Accelerated Course Archives Using Condor with Blackboard Sam Hoover, IT Systems Architect Matt Garrett, System Administrator.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
Jia Yao Director: Vishwani D. Agrawal High Performance Compute Cluster April 13,
Information Technology Center Introduction to High Performance Computing at KFUPM.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Academic and Research Technology (A&RT)
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
 Contents 1.Introduction about operating system. 2. What is 32 bit and 64 bit operating system. 3. File systems. 4. Minimum requirement for Windows 7.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
ISG We build general capability Purpose After this tutorial, you should: Be comfortable submitting work to the batch queuing system of olympus and be familiar.
Utilizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover 1 Project Blackbird Computing,
JGI/NERSC New Hardware Training Kirsten Fagnan, Seung-Jin Sul January 10, 2013.
Introduction to UNIX/Linux Exercises Dan Stanzione.
Types of Operating System
Introduction to HP LoadRunner Getting Familiar with LoadRunner >>>>>>>>>>>>>>>>>>>>>>
Chapter 3: Operating-System Structures System Components Operating System Services System Calls System Programs System Structure Virtual Machines System.
Introduction to HPC resources for BCB 660 Nirav Merchant
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
MaterialsHub - A hub for computational materials science and tools.  MaterialsHub aims to provide an online platform for computational materials science.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Common Practices for Managing Small HPC Clusters Supercomputing 12
Experiences with a HTCondor pool: Prepare to be underwhelmed C. J. Lingwood, Lancaster University CCB (The Condor Connection Broker) – Dan Bradley
SSS Test Results Scalability, Durability, Anomalies Todd Kordenbrock Technology Consultant Scalable Computing Division Sandia is a multiprogram.
How to get started on cees Mandy SEP Style. Resources Cees-clusters SEP-reserved disk20TB SEP reserved node35 (currently 25) Default max node149 (8 cores.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Linux & Shell Scripting Small Group Lecture 3 How to Learn to Code Workshop group/ Erin.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
M. Schott (CERN) Page 1 CERN Group Tutorials CAT Tier-3 Tutorial October 2009.
Weekly Report By: Devin Trejo Week of June 14, 2015-> June 20, 2015.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
HUBbub 2013: Developing hub tools that submit HPC jobs Rob Campbell Purdue University Thursday, September 5, 2013.
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Cloud Computing project NSYSU Sec. 1 Demo. NSYSU EE IT_LAB2 Outline  Our system’s architecture  Flow chart of the hadoop’s job(web crawler) working.
 System Requirements are the prerequisites needed in order for a software or any other resources to execute efficiently.  Most software defines two.
Computer Performance. Hard Drive - HDD Stores your files, programs, and information. If it gets full, you can’t save any more. Measured in bytes (KB,
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
IPPP Grid Cluster Phil Roffe David Ambrose-Griffith.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
GRID COMPUTING.
Welcome to Indiana University Clusters
HPC usage and software packages
Welcome to Indiana University Clusters
GWE Core Grid Wizard Enterprise (
ASU Saguaro 09/16/2016 Jung Hyun Kim.
MaterialsHub - A hub for computational materials science and tools.
CommLab PC Cluster (Ubuntu OS version)
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
High Performance Computing in Bioinformatics
Introduction to High Performance Computing Using Sapelo2 at GACRC
LO2 – Understand Computer Software
Working in The IITJ HPC System
Presentation transcript:

ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing Center MISSION 2.0 Training, Dec 11 th,

ISG We build general capability This presentation will answer the following questions… What is Olympus? Why do I need Olympus? How do I and my group get access to Olympus? How do I use Olympus? What will I learn in the hands-on training? 2

ISG We build general capability What is Olympus? 3 Olympus is a compute cluster hosted by ISG to provide CPU time for MIDAS network research.

ISG We build general capability RAM Disk What is Olympus? Core – a single processing unit capable of performing computations. Multicore Processor – a unit that contains two or more independent cores Socket – place where a multicore processor is plugged in and connects to other resources in the computer Node – collection of sockets and share a common memory space Cluster – a collection of nodes connected through an external network 4 What is a cluster?

ISG We build general capability What is Olympus? It is a cluster (1536 cores, 7.5 TB of RAM) 24 nodes (64 cores each) 4 sockets AMD multicore processor 16 cores (1.4 GHz) 512 GB of RAM (8 Nodes) 256 GB of RAM (16 Nodes) Local 3TB hard disk Nodes connected to each other by Gigabit Ethernet Shared file System (2TB) Upgraded Q with ~50 TB file system Software available: Python, R, gcc, openmpi Simulators available: FRED and FluTE Additional software installed on request 5

ISG We build general capability Why do I need Olympus? I can do more work, more quickly. More capacity allows for more simulations 1 million FRED county level influenza simulations per day 100,000 FRED state level influenza simulations per day I can do larger work. Large memory nodes allow bigger simulations 28 FRED national level influenza simulations per day I can work more efficiently and effectively. Queuing system allows setting up large batches of runs through scripting rather than manual input creation Support from PSC/ISG staff to help you work more efficiently. 6

ISG We build general capability How do I get access to Olympus? Can I get an account? Anyone doing MIDAS related research can get an account on Olympus. How? Request an account through the MIDAS Portal I have an account, now what do I do? You can access the machine through ssh at the address olympus.psc.edu. What does an account get me? Home directory for storing files Ability to submit jobs to the machine Support if you run into trouble and/or need software installed Consultation for working more efficiently and effectively 7

ISG We build general capability How do I use Olympus? When you log onto the machine, you will be on the head node. A limited machine that allows you to interact with the full cluster. What should I do on the head node? Manage your files Setup work that you plan to submit to the machine How I do I actually run a job on Olympus? Jobs will need to be submitted to a queuing system which will put the job in line for execution when resources are available. 8

ISG We build general capability ssh How do I use Olympus? What is a queuing system? Software allowing users to submit jobs on a multiuser machine by scheduling jobs on the machine and running when the resources are available. Olympus runs the PBS / TORQUE queuing system ( as standard in the world of supercomputing) How do I use the queuing system? Prepare your job on the head node Create a job submission script Defines the job to be accomplished Defines the resources needed to accomplish the job (e.g. how many cores do I need) Submit the job to the queue Check status of the job (queued, running, done) The job will run when the resources are available When finished, collect results on the head node I want to run simulations. I am going to create a whole lot of inputs to run. I need 32 cores, 128 GB of RAM and 2 hours of time. Head Node olympus.psc.edu Olympus Cluster PBS / TORQUE Queuing System WORK Inputs Queued Running Output Wow, I got a lot of work done! I am going to go get coffee. Done 9

ISG We build general capability What will I learn in the hands- on training at noon? How do I log on to Olympus? How do I create a submission script? How do I submit and monitor work through the queuing system? What more do I need to know? How do I run interactively on Olympus? How do I effectively use the file systems on Olympus? 10