Publishing applications on the web via the Easa Portal and integrating the Sun Grid Engine Publishing applications on the web via the Easa Portal and integrating.

Slides:



Advertisements
Similar presentations
White Rose Grid Infrastructure Overview Chris Cartledge Deputy Director Corporate Information and Computing Services, The University of Sheffield
Advertisements

Global Grid Access Cezary Mazurek, PSNC. Cezary Mazurek, PSNC, Enable access to global grid, Supercomputing 2003, Phoenix, AZ 2 Agenda Introduction PROGRESS.
Beowulf Supercomputer System Lee, Jung won CS843.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Information Technology Center Introduction to High Performance Computing at KFUPM.
LUNARC, Lund UniversityLSCS 2002 Transparent access to finite element applications using grid and web technology J. Lindemann P.A. Wernberg and G. Sandberg.
Monitoring and performance measurement in Production Grid Environments David Wallom.
Network+ Guide to Networks, Fourth Edition Chapter 10 Netware-Based Networking.
OxGrid, A Campus Grid for the University of Oxford Dr. David Wallom.
6/2/20071 Grid Computing Sun Grid Engine (SGE) Manoj Katwal.
The Cactus Portal A Case Study in Grid Portal Development Michael Paul Russell Dept of Computer Science The University of Chicago
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Introduction to Grid Computing with High Performance Computing Mike Griffiths White Rose Grid e-Science Centre of Excellence.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Project Implementation for COSC 5050 Distributed Database Applications Lab1.
Internet GIS. A vast network connecting computers throughout the world Computers on the Internet are physically connected Computers on the Internet use.
Welcome to the Minnesota SharePoint User Group. Introductions / Overview Project Tracking / Management / Collaboration via SharePoint Multiple Audiences.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
6/1/2001 Supplementing Aleph Reports Using The Crystal Reports Web Component Server Presented by Bob Gerrity Head.
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
Grid Information Systems. Two grid information problems Two problems  Monitoring  Discovery We can use similar techniques for both.
Using the WDK for Windows Logo and Signature Testing Craig Rowland Program Manager Windows Driver Kits Microsoft Corporation.
Gilbert Thomas Grid Computing & Sun Grid Engine “Basic Concepts”
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
Introduction To Computer System
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
Software Architecture
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Chapter 34 Java Technology for Active Web Documents methods used to provide continuous Web updates to browser – Server push – Active documents.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
The PROGRESS Grid Service Provider Maciej Bogdański Portals & Portlets 2003 Edinburgh, July 14th-17th.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
PROGRESS: ICCS'2003 GRID SERVICE PROVIDER: How to improve flexibility of grid user interfaces? Michał Kosiedowski.
NGS Innovation Forum, Manchester4 th November 2008 Condor and the NGS John Kewley NGS Support Centre Manager.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
CCS Overview Rene Salmon Center for Computational Science.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
1 Media Grid Initiative By A/Prof. Bu-Sung Lee, Francis Nanyang Technological University.
Grid-Powered Scientific & Engineering Applications Ho Quoc Thuan INSTITUTE OF HIGH PERFORMANCE COMPUTING.
6/1/2001 Supplementing Aleph Reports Using The Crystal Reports Web Component Server Presented by Bob Gerrity Head.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
NEES Cyberinfrastructure Center at the San Diego Supercomputer Center, UCSD George E. Brown, Jr. Network for Earthquake Engineering Simulation NEES TeraGrid.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
International Symposium on Grid Computing (ISGC-07), Taipei - March 26-29, 2007 Of 16 1 A Novel Grid Resource Broker Cum Meta Scheduler - Asvija B System.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
GraDS MacroGrid Carl Kesselman USC/Information Sciences Institute.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre
PROGRESS: GEW'2003 Using Resources of Multiple Grids with the Grid Service Provider Michał Kosiedowski.
NGS computation services: APIs and.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Advanced Computing Facility Introduction
Clouds , Grids and Clusters
White Rose Grid Infrastructure Overview
Grid Computing.
NGS computation services: APIs and Parallel Jobs
University of Technology
Introduction to Cloud Computing
Unit 27: Network Operating Systems
H2020 EU PROJECT | Topic SC1-DTH | GA:
Presentation transcript:

Publishing applications on the web via the Easa Portal and integrating the Sun Grid Engine Publishing applications on the web via the Easa Portal and integrating the Sun Grid Engine By Michael Griffiths & Deniz Savas CiCS Dept. Sheffield University Sept 2007

Sheffield is in South Yorkshire, England

Sheffield University- facts Established in academic departments in 7 faculties Number of Undergraduate Students: 25,500 Number of Post Graduate/Research Students: 5,600 Number of International Students : 3,100

‘iceberg’ the HPC Cluster at the Computer Centre AMD Opteron based, supplied by Sun Microsystems. Processors: 320 ( 160 of these are designated to the Physics Dept. for the PP project ) Performance: 300GFLOPs Main Memory: 800GB User filestore: 9TB Temporary disk space: 10TB Physical size: 8 racks Power usage: 50KW

‘iceberg’ cluster hardware components 160 general-purpose-use cpu’s; 80 of these are in dual-core configuration with 2 GBytes of memory each. ( V20 Model ) (i.e 40 boxes with 2 cpus + 4 GBytes ) 80 are in quad-core configurations with 4 GBytes memory each. ( V40 Model ) ( i.e 20 boxes with 4 cpus + 16 GBytes ) These are also connected via a Myrinet Switch at 2Gbps connection speed. IPMI Service Processors : Each box contains a service processor with separate network interface for remote monitoring and control. Inside a V20

Iceberg Cluster Configuration HEAD NODE Worker node 7 Service Proc 7 Worker node 1 Service Proc 1 Worker node 6 Service Proc 6 Worker node n Service Proc n Worker node 56 Service Proc n Worker node 2 Service Proc 2 Worker node n Service Proc n Worker node 57 Service Proc 57 Worker node 4 Service Proc 4 Worker node 9 Service Proc 9 Worker node n Service Proc n Worker node 59 Service Proc 59 Worker node 3 Service Proc 1 Worker node 8 Service Proc 8 Worker node n Service Proc n Worker node 58 Service Proc 58 Worker node 5 Service Proc 5 Worker node 10 Service Proc 10 Worker node n Service Proc n Worker node 60 Service Proc 60 Eth0 Eth1 nfs mounted onto Worker nodes Shared file store All remote access License server Myranet Connected Workers

Iceberg Cluster Configuration HEAD NODE Worker node 7 Service Proc n Worker node 1 Service Proc 1 Worker node 6 Service Proc n Worker node n Service Proc n Worker node 56 Service Proc n Worker node 2 Service Proc 2 Worker node n Service Proc n Worker node 57 Service Proc n Worker node 4 Service Proc 1 Worker node 9 Service Proc n Worker node n Service Proc n Worker node 59 Service Proc n Worker node 3 Service Proc 1 Worker node 8 Service Proc n Worker node n Service Proc n Worker node 58 Service Proc n Worker node 5 Service Proc 1 Worker node 10 Service Proc n Worker node n Service Proc n Worker node 60 Service Proc n Eth0 nfs mounted onto Worker nodes Shared file store Eth1

Iceberg Cluster Configuration Worker node 7 Service Proc n Worker node 1 Service Proc 1 Worker node 6 Service Proc n Worker node n Service Proc n Worker node 56 Service Proc n Worker node 2 Service Proc 2 Worker node n Service Proc n Worker node 57 Service Proc n Worker node 4 Service Proc 1 Worker node 9 Service Proc n Worker node n Service Proc n Worker node 59 Service Proc n Worker node 3 Service Proc 1 Worker node 8 Service Proc n Worker node n Service Proc n Worker node 58 Service Proc n Worker node 5 Service Proc 1 Worker node 10 Service Proc n Worker node n Service Proc n Worker node 60 Service Proc n HEAD NODE

White Rose Grid YHMAN Network

Grid & HPC applications development tools Development Fortran77,90, C, C++, Java compilers MPI / MPICH-gm OpenMP Nag Mk 20, 21 ACML Grid Sun Grid Engine Globus (via gpt 3.0) SRB s-client tools

Using the White Rose Grid Application Portal

Features and Capabilities Web accessible management and execution of applications Provides a service for rapid authoring and publication of custom applications Easily integrate multiple heterogeneous resources

Potential benefits of an applications portal More efficient use of resources Ease of use Familiar GUI interface Capturing of expert knowledge Better presentation of legacy software

Potential Development Building Expert Systems Allowing novice expert to take advantage of parallel HPC resources Providing HPC services over the grid HPC centres collaborating with each other without having to provide individual usernames, file-storage etc to remote users.

WRG – Application Portal Based on EASA Three Usage Modes Users Run applications Have storage space Review old results Authors Build and publish applications Administrators

Using Accessing Managing Applications Workspace Results Help

Using:Accessing Start up a web browser & a.html Login using provided user name and password

Using:Help Select Help and Support tab to register Apply to Admin for an account Apply to authors to register applications

Using:Managing Installing a client Setting password Setting Mode user/author

Using:Applications View and select available applications

Running An Application

User Interface

Using:Workspace Storage for uploaded files and old job files

Using:Results Check results View job progress Export to spreadsheet

Using: Results Viewing Results

Using:Help Documentation Contacts

Conclusions Disadvantages Thick Client, License costs Advantages Rapid publication Enable virtualization of HPC resources Make applications available to broader community, become application focused Effective on a network with low bandwidth Make applications available to collaboration partners over the internet and outside own organisation

Demonstration Applications Developed for EASA Demonstration of Metascheduling Across White Rose Grid Monitoring of usage across White Rose Grid Running Applications on the local cluster Fluent Ansys Generic Matlab and Scilab applications

Metascheduler Demonstration:Background Enable utilisation of resources across White Rose Grid Exploit use of task arrays Job submission is seamless Demonstration uses a generic scilab application that runs on any of the White Rose Grid Nodes Simplistic, but; effective, manageable and sustainable

Metascheduler Demonstration: Method Query and Compare job queues for WRG nodes qstat –g c Use slots available and total number of slots to generate weights for different queues Compare weights for all queues on different nodes and use to select node Use standard EASA job submission technique to submit job to selected node EASA does not know about clusters Special easaqsub submits job to sge, monitors job status will remove job if wait time exceeded, easaqsub job monitor has completed EASA knows that EASA compute task has completed

Metascheduler Demonstration: Running Scilab User provides scilab scriptfile Required resource file e.g. datafiles or files for scilab library routines Can provide zipped bundle of scilab resources Set job submission information and then submit job

Metascheduler Demonstration: Job Submission Provide jobname and job description Information used for metascheduling Jobtime (hours) Waittime (hours) Number of tasks (for job array) Submission method Use metascheduling Select a particular node

Metascheduler Demonstration : Further Developments Current method successful! Correctly selects clusters and improves turnaround for scilab compute tasks Current pattern can be extended to other EASA applications Provide distributed storage across White rose Grid Develop metascheduling strategy introduce greater dependency on user job requirements for node selection Exploit other metascheduling systems e.g. SGE transfer queues, CONDOR-G THE END