Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.

Slides:



Advertisements
Similar presentations
E. Ferro, CNAF, aprile LCFG: testbed upgrade Enrico Ferro INFN-LNL.
Advertisements

All Hands MeetingD. Colling, Imperial College London, for the GridPP Project 1 Hans Hoffman has described the scale of the problems that we are facing,
Andrew McNab - Manchester HEP - 10 May 2002 UK Testbed Deployment Aim of this talk is to the answer the questions: –What are other sites doing? –What are.
Andrew McNabTestbed / HTTPS, GridPP6, 30 Jan 2003Slide 1 UK Testbed Status Andrew McNab High Energy Physics University of Manchester.
Status Report University of Bristol 3 rd GridPP Collaboration Meeting 14/15 February, 2002Marc Kelly University of Bristol 1 Marc Kelly University of Bristol.
ATLAS/LHCb GANGA DEVELOPMENT Introduction Requirements Architecture and design Interfacing to the Grid Ganga prototyping A. Soroko (Oxford), K. Harrison.
Andrew McNab - Manchester HEP - 24 May 2001 WorkGroup H: Software Support Both middleware and application support Installation tools and expertise Communication.
12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
Bob Jones – Project Architecture - 1 March n° 1 Information & Monitoring Services Antony Wilson WP3, RAL
Presenter Name Facility Name EDG Testbed Status Moving to Testbed Two.
Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
Andrew McNab - Manchester HEP - 31 January 2002 Testbed Release in the UK Integration Team UK deployment TB1 Job Lifecycle VO: Authorisation VO: GIIS and.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
EU 2nd Year Review – Jan – Title – n° 1 WP1 Speaker name (Speaker function and WP ) Presentation address e.g.
J. Hanton - P. Herquet - F. Lequeux - A. Romeyer1 CONDOR-G Installation July 2004 : one independent PC for Grid FTP as a client to UCL August 2004 : complete.
INFN - Ferrara BaBarGrid Meeting SPGrid Efforts in Italy BaBar Collaboration Meeting - SLAC December 11, 2002 Enrica Antonioli - Paolo Veronesi.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Andrew McNab - Manchester HEP - 29/30 March 2001 gridmapdir patch Overview of the problem Constraints from local systems Outline of how it works How to.
Andrew McNab - EDG Access Control - 14 Jan 2003 EU DataGrid security with GSI and Globus Andrew McNab University of Manchester
The EDG Testbed Introduction and Setup The European DataGrid Project Team
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
Andrew McNab - Manchester HEP - 6 November Old version of website was maintained from Unix command line => needed (gsi)ssh access.
Andrew McNab - Manchester HEP - 22 April 2002 UK Rollout and Support Plan Aim of this talk is to the answer question “As a site admin, what are the steps.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
The EDG Testbed Deployment Details The European DataGrid Project
Andrew McNab - Manchester HEP - 26 June 2001 WG-H / Support status Packaging / RPM’s UK + EU DG CA’s central grid-users file grid “ping”
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
1 BIG FARMS AND THE GRID Job Submission and Monitoring issues ATF Meeting, 20/06/03 Sergio Andreozzi.
Computational grids and grids projects DSS,
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
DataGrid WP1 Massimo Sgaravatto INFN Padova. WP1 (Grid Workload Management) Objective of the first DataGrid workpackage is (according to the project "Technical.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Grid job submission using HTCondor Andrew Lahiff.
Andrew McNabETF Firewall Meeting, NeSC, 5 Nov 2002Slide 1 Firewall issues for Globus 2 and EDG Andrew McNab High Energy Physics University of Manchester.
EDG Testbed installation and configuration with LCFGng Maite Barroso - WP4
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
05/29/2002Flavia Donno, INFN-Pisa1 Packaging and distribution issues Flavia Donno, INFN-Pisa EDG/WP8 EDT/WP4 joint meeting, 29 May 2002.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
London Tier 2 Status Report GridPP 11, Liverpool, 15 September 2004 Ben Waugh on behalf of Owen Maroney.
Andrew McNab - Manchester HEP - 11 May 2001 Packaging / installation Ready to take globus from prerelease to release. Alex has prepared GSI openssh.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
University of Bristol 5th GridPP Collaboration Meeting 16/17 September, 2002Owen Maroney University of Bristol 1 Testbed Site –EDG 1.2 –LCFG GridPP Replica.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
Portal Update Plan Ashok Adiga (512)
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Andrew McNab - Manchester HEP - 17 September 2002 UK Testbed Deployment Aim of this talk is to the answer the questions: –“How much of the Testbed has.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Andrew McNabGrid in 2002, Manchester HEP, 7 Jan 2003Slide 1 Grid Work in 2002 Andrew McNab High Energy Physics University of Manchester.
LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
Andrew McNab - Security issues - 17 May 2002 WP6 Security Issues (some personal observations from a WP6 and sysadmin perspective) Andrew McNab, University.
Andrew McNab - Security issues - 4 Mar 2002 Security issues for TB1+ (some personal observations from a WP6 and sysadmin perspective) Andrew McNab, University.
Andrew McNab - Dynamic Accounts - 2 July 2002 Dynamic Accounts in TB1.3 What we could do with what we’ve got now... Andrew McNab, University of Manchester.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
The EDG Testbed The European DataGrid Project Team
LHCb Grid MeetingLiverpool, UK GRID Activities Glenn Patrick Not particularly knowledgeable-just based on attending 3 meetings.  UK-HEP.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Software Management Workshop Steve Traylen. Software Management(WG5) The aim of the working group is to look at deficiencies in deployed and upcoming.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
The EDG Testbed Deployment Details
UK GridPP Tier-1/A Centre at CLRC
Artem Trunov and EKP team EPK – Uni Karlsruhe
The EU DataGrid Fabric Management Services
Presentation transcript:

Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed. Done with a handful of modifications to the Testbed site and to the existing farms. This talks describes what we did and how you can do it too...

Andrew McNab - Manchester HEP - 17 September 2002 Farms at Manchester HEP BaBar 80 * 0.8GHz GridFarm 16 * 1.0GHz DZero / Atlas 60 * 1.5GHz

Andrew McNab - Manchester HEP - 17 September 2002 The problem We want to make existing farms available on the Testbed. But we dont want to massively reconfigure/reinstall farms –theyre in production so need to be kept stable –they are already configured the way their owners need We might want to keep reinstalling as EDG software is updated. –this is labour intensive unless we install from scratch with LCFG install –dont want to have to make many manual changes to CE etc every time we install/upgrade Solution that has been mentioned several times is to have a standard EDG Testbed Site as a front end to the Existing Farm So want to find the minimal set of changes to Farm and Testbed Site that will put the Farm on the Testbed.

Andrew McNab - Manchester HEP - 17 September 2002 Standard Testbed Site SE LCFG WN CEWN /home /flatfiles PBS Node PBS Server PBS All elements installed from LCFG server Computing Element shares /home directories by NFS Storage Element shares /flatfiles with data by NFS PBS Server on CE talks to PBS on Worker Nodes.

Andrew McNab - Manchester HEP - 17 September 2002 What we want SE LCFG WN CEWN /home /flatfiles PBS Node PBS Server PBS Server PBS Node PBS Node Grid Farm / Testbed Site BaBar or DZero/Atlas Farm PBS qsub

Andrew McNab - Manchester HEP - 17 September 2002 Reconfigure Existing Farm PBS Server must allow access from CE, but only for the right users. –Add CE to list of valid job submission clients (eg in hosts.equiv) –Create special queue (bfq or dfq) for Testbed jobs. –Limit queues so desired pool of accounts (eg atlas001 etc) can submit jobs to the bfq/dfq but other queues/pools forbidden. PBS Nodes need access to pool accounts, home directories on CE, and /flatfiles area on SE. –If already using NFS automount, then easy to add /home on CE and /flatfiles on SE (eg as /nfs/gf-home and /nfs/gf-flatfiles) –Add pool accounts to /etc/passwd (or NIS) –Make symbolic links in /home to automount CE /home directories.

Andrew McNab - Manchester HEP - 17 September 2002 Software on PBS Nodes For current EDG job submissions to work, need to install globus-url-copy RPMs on PBS Nodes. –PBS Nodes currently need to make an outgoing gridftp –connections to Resource Broker. –GridFTP possible with NAT, but difficult. Other middleware RPMs will be needed if also intending to manipulate SE and RC during jobs. For use with EDG Testbed, should also install relevant application RPMs

Andrew McNab - Manchester HEP - 17 September 2002 Changes to Testbed Site Have attempted to minimise changes: –easier to document and support –easier to maintain as EDG software changes Basic philosophy: modify EDG scripts to make remote qsub and qstat calls to PBS Server machines on the farms. Only need to edit 3 scripts on the CE –/opt/globus/libexec/globus-script-pbs-queue –/opt/edg/info/mds/sbin/skel/ce-globus.skel –/opt/edg/info/mds/bin/ce-pbs Create grid-mapfile and ce-static.ldif for each queue. Include farm queue and PBS nodes in LCFG site-cfg.h

Andrew McNab - Manchester HEP - 17 September 2002 New behaviour Modified ce-pbs queries PBS Server using remote qstat Publishes edited grid-mapfile listing only the right users. Jobs can be submitted using Resource Broker, based on published information. When received by CE, globus-script-pbs-queue submits job to remote PBS Server EDG Globus jobmanager on CE monitors job status via remote qstat and transmits to Logging as normal. Job runs on PBS Node with access to pool account /home Job completes and returns files to RB via gridftp

Andrew McNab - Manchester HEP - 17 September 2002 Example logs Three jobmanagers visible to GridPP MDS and RB: –gf18.hep.man.ac.uk:2119/jobmanager-pbs-gfq(Grid Farm/Testbed) –gf18.hep.man.ac.uk:2119/jobmanager-pbs-dfq(DZero/Atlas farm) –gf18.hep.man.ac.uk:2119/jobmanager-pbs-bfq(BaBar farm) Different operating system, grid-mapfile lists of users etc for each queue. Can submit job to RB and have it matchmake the requirements –including dynamic properties like free nodes Example log shows submitting a job from UI at RAL via RB at IC, which decides which farm at Manchester matches and sends the job there.

Andrew McNab - Manchester HEP - 17 September 2002 Applying this to other sites This recipe being written up for With current EDG release, the PBS Nodes need outgoing direct internet access (not NAT.) You need to be able to make minor changes to PBS Server permissions, NFS mounts etc as described. You should have some (3?) dedicated Testbed machines, or add it to an existing GridPP/EDG Testbed setup. –We use Microdirect.co.uk boxes at 1.5GHz/256MB/40GB box for £250 …. If you dont use an EDG-supported batch system (PBS etc), you need to modify ce-pbs and globus-script-pbs-* scripts to use your job submission commands.

Andrew McNab - Manchester HEP - 17 September 2002 Summary Its not at all difficult to access existing PBS farms via an EDG Testbed site. –include CE + SE in NFS and PBS configuration of farm –include pool accounts in farms passwd file –enforce security by account pools Only need to modify a handful of files on the Testbed CE. Should be relatively straightforward to apply this to other batch queue systems even if you dont use PBS. Weve demonstrated putting our 150 * ~1 GHz nodes on the current Testbed and submitting jobs via GridPP RB You can too.