17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.

Slides:



Advertisements
Similar presentations
Parallel ISDS Chris Hans 29 November 2004.
Advertisements

Publishing applications on the web via the Easa Portal and integrating the Sun Grid Engine Publishing applications on the web via the Easa Portal and integrating.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
CCPR Workshop Lexis Cluster Introduction October 19, 2007 David Ash.
Southgreen HPC system Concepts Cluster : compute farm i.e. a collection of compute servers that can be shared and accessed through a single “portal”
PBS Job Management and Taskfarming Joachim Wagner
ISG We build general capability Job Submission on the Olympus Cluster J. DePasse; S. Brown, PhD; T. Maiden Pittsburgh Supercomputing Center Public Health.
Information Technology Center Introduction to High Performance Computing at KFUPM.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Job Submission on WestGrid Feb on Access Grid.
LUNARC, Lund UniversityLSCS 2002 Transparent access to finite element applications using grid and web technology J. Lindemann P.A. Wernberg and G. Sandberg.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
HPCC Mid-Morning Break Powertools Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Research
HPCC Mid-Morning Break Interactive High Performance Computing Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
ISG We build general capability Purpose After this tutorial, you should: Be comfortable submitting work to the batch queuing system of olympus and be familiar.
Communicating with Users about HTCondor and High Throughput Computing Lauren Michael, Research Computing Facilitator HTCondor Week 2015.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
Introduction to UNIX/Linux Exercises Dan Stanzione.
DIANE Overview Germán Carrera, Alfredo Solano (CNB/CSIC) EMBRACE COURSE Monday 19th of February to Friday 23th. CNB-CSIC Madrid.
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
Building service testbeds on FIRE D5.2.5 Virtual Cluster on Federated Cloud Demonstration Kit August 2012 Version 1.0 Copyright © 2012 CESGA. All rights.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Introduction to the HPCC Jim Leikert System Administrator High Performance Computing Center.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Debugging and Profiling GMAO Models with Allinea’s DDT/MAP Georgios Britzolakis April 30, 2015.
HYDRA: Using Windows Desktop Systems in Distributed Parallel Computing Arvind Gopu, Douglas Grover, David Hart, Richard Repasky, Joseph Rinkovsky, Steve.
INVITATION TO COMPUTER SCIENCE, JAVA VERSION, THIRD EDITION Chapter 6: An Introduction to System Software and Virtual Machines.
Network Monitoring System for the UNIX Lab Bradley Kita Capstone Project Mentor: Dr C. David Shaffer Fall 2004/Spring 2005.
Network Queuing System (NQS). Controls batch queues Only on Cray SV1 Presently 8 queues available for general use and one queue for the Cray analyst.
HPC for Statistics Grad Students. A Cluster Not just a bunch of computers Linked CPUs managed by queuing software – Cluster – Node – CPU.
CCS Overview Rene Salmon Center for Computational Science.
Operating Systems Lesson Objective: Understanding the functions of an operating system. Learning Outcome: Answer some basic questions on operating systems.
NEES Cyberinfrastructure Center at the San Diego Supercomputer Center, UCSD George E. Brown, Jr. Network for Earthquake Engineering Simulation NEES TeraGrid.
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat.
Unix Machine In Computer Science for Teaching Cliff Zou Spring 2015.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
The Gateway Computational Web Portal Marlon Pierce Indiana University March 15, 2002.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
Portable Batch System – Definition and 3 Primary Roles Definition: PBS is a distributed workload management system. It handles the management and monitoring.
Chapter 4 Software. Introduction Program: is a set of sequence instructions that tell the computer what to do. Software: is a collection of programs,
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Auburn University
Welcome to Indiana University Clusters
HPC usage and software packages
Welcome to Indiana University Clusters
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
CommLab PC Cluster (Ubuntu OS version)
Rui Wu, Jose Painumkal, Sergiu M. Dascalu, Frederick C. Harris, Jr
Compiling and Job Submission
Advanced Computing Facility Introduction
Quick Tutorial on MPICH for NIC-Cluster
Using and Building Infrastructure Clouds for Science
Working in The IITJ HPC System
Presentation transcript:

17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin

17-April-2007 Outline  What is the HPC?  Where did it come from?  How can you get an account on hpc.mnsu.edu?  How can you use it for your research?  Where do you go from here?

17-April-2007 What is the HPC?  Many AMD Opteron Computers (nodes) in a rack  Connected by a high- speed network  In the IT Services Secure area (third floor of the library)  All nodes run linux 

17-April-2007 What is the HPC?  Head node has 8GB RAM; 7.4 TB of Disk  Head node is for doing administrative work and starting long jobs  The 34 Worker nodes are for doing long computations  Each worker has 8GB RAM; 80 GB Hard Disk; 2 dual-core AMD Opteron Head Node Worker 1 …Worker 34

17-April-2007 What is the HPC?  Software Installed: <GNU languages: C/C++ (gcc/g++), Fortran (gfortran) <Message Passing Interface library OpenMPI  Software soon to be installed: <MATLAB <Fluent <Portland Group Fortran and C/C++ <IMSL  is “local delivery only”

17-April-2007 Where did it come from?  National Science Foundation Grant <MRI Program (Major Research Instrumentation) <$140,000 <Institutional Equipment funds upgraded machine by adding five nodes  PIs: Patrick Tebbe, Rebecca Bates, David Haglin  Proposal focused on a college-wide need for HPC  Vendor: PSSC Labs, Inc.

17-April-2007 How can you get an account?  We must submit a final report to NSF after July 31, 2009  Part of the final report must include how much it was used within CSET (and within MSU).  We need to track usage (research projects).  To get an account, send an to with information as < <Your students can get accounts too!  We are very interested in knowing about publications you obtain as a result of using hpc.mnsu.edu.

17-April-2007 Your Research  Okay, so you got an account.  Now What?

17-April-2007 Your Research  Learning to use HPC.  Learning to use the OpenPBS/Torque job queuing software.  Learning to “design” your usage.  Tutorials will be maintained at

17-April-2007 Your Research  Connect to hpc.mnsu.edu (head node) using ssh <ssh on unix <PuTTY or SSH Windows Client (IT Services) <Firewall is pretty tight, may need to request a new opening in the firewall from your location  Line-mode (command-line) interface  Basic unix commands: <

17-April-2007 Your Research  Disks on hpc:

17-April-2007 Your Research  Using OpenPBS/Torque job queuing software: <qstat -- Inspect current job queue <qsub -- Add a new job to the queue <qdel-- Delete one of your jobs from the Q <pbsmon.py-- See the state of the entire machine <xpbsmon-- Uses X11 to display machine state <firefox localhost/ganglia  Detailed information available at: < manual.shtmlhttp:// manual.shtml

17-April-2007 Your Research  Designing your usage. <Assume you have a program you want to run for different parameter values of 1 through 1000 <Ex: $ myProgram -p1 $ myProgram -p2. $ myProgram -p1000

17-April-2007 Your Research  Create 1000 “start scripts” to queue 1000 jobs to the master queue.  Start your jobs and monitor their progress  Combine results when they are all done.  Organize experiments/runs in folders  Use scripting languages such as python to generate start scripts.

17-April-2007 Your Research  Input and Output for your jobs: <Your script will start on a worker node <You can log in to a worker node to see filesystem:  ssh n04  df <Standard Output and Standard Error are separate <Files are written alongside your script when jobs completes <No way to monitor progress of your computation

17-April-2007 Your Research  Sample script to run from 501 to 505:

17-April-2007 Where do you go from here?  is a communication portal  Find colleagues who can help  Learn more about the capabilities: <New software <Parallel programming (MPI) <Parallel libraries: e.g., ScaLAPACK.  Keep this machine computing fast  Other ideas?