The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures 2007 17 th December 2007.

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Advertisements

Liverpool HEP – Site Report May 2007 John Bland, Robert Fay.
A couple of slides on RAL PPD Chris Brew CCLRC - RAL - SPBU - PPD.
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
17th October 2013Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Presented by Jacob Wilson SharePoint Practice Lead Bross Group 1.
SouthGrid Status Pete Gronbech: 4 th September 2008 GridPP 21 Swansea.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
RAL PPD Site Update and other odds and ends Chris Brew.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
RAL PPD Computing A tier 2, a tier 3 and a load of other stuff Rob Harper, June 2011.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Steve Traylen PPD Rutherford Lab Grid Operations PPD Christmas Lectures Steve Traylen RAL Tier1 Grid Deployment
13th October 2011Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
14th October 2010Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and SouthGrid Technical Co-ordinator.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
11th October 2012Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
EGEE is a project funded by the European Union under contract IST VO box: Experiment requirements and LCG prototype Operations.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
BaBar Cluster Had been unstable mainly because of failing disks Very few (
Scientific Computing in PPD and other odds and ends Chris Brew.
RALPP Site Report HEP Sys Man, 11 th May 2012 Rob Harper.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
Patrick Gartung 1 CMS 101 Mar 2007 Introduction to the User Analysis Facility (UAF) Patrick Gartung - Fermilab.
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
Glasgow Site Report (Group Computing)
SAM at CCIN2P3 configuration issues
NGS Oracle Service.
Printer Admin Print Job Manager
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
Presentation transcript:

The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007

2 17/12/2007 PPD Christmas Lectures 2007 Contents What’s Tier 2 or Tier 3? Tier 2 and Tier 3 Hardware Staff Changes over the last year Some (new) details of the Tier 3 Service

3 17/12/2007 PPD Christmas Lectures 2007 Tier 2 vs. Tier 3 The RAL PPD Tier 2 is the grid resources committed to GridPP/WLCG/EGEE as part of the SouthGrid Distributed Tier 2 –The Grid Batch Farm, the Storage Element The RAL PPD Tier 3 is the local user cluster available to the department –The User Interfaces, the local file servers

4 17/12/2007 PPD Christmas Lectures 2007 The Tier 2 Hardware 384 Batch CPU cores –48 x Intel 2.8GHz PIV –208 x AMD 2.0GHz Opteron 270 –128 x Intel 2.0GHz Woodcrest TeraBytes of Disk Space in dCache –8 x 10TB servers –13 x 6TB servers Various middleware and infrastructure nodes 10Gb/s Link to Site Backbone (and so Tier 1) New

5 17/12/2007 PPD Christmas Lectures 2007 The Tier 3 Hardware 3 Disk servers: –home and software servers Pair of 1GB RAID servers Home and software areas cross sync’d daily –Misc server 6.4TB RAID server Hosts Scratch, installation and other miscellaneous areas

6 17/12/2007 PPD Christmas Lectures 2007 Tier 3 Hardware Continued 8 User Interfaces –heplnx101, 102 general SL4 –heplnx103 general SL3 –heplnx104 CMS SL3 PhEDEx –heplnx105, 106 Atlas SL3 –heplnx107 CMS SL4 –heplnx108 LHCb SL4 All * User interfaces upgraded to faster hardware with more memory

7 17/12/2007 PPD Christmas Lectures 2007

8 17/12/2007 PPD Christmas Lectures 2007 Support Staff Currently well below compliment –50% of Chris Brew In the process of recruiting 2 new system administrators With extra effort we will be able to do more things

Changes over the last year

10 17/12/2007 PPD Christmas Lectures 2007 SL4 Migration All batch capacity, many backend servers plus half the front ends upgraded to SL4 Remaining front ends will be upgraded when SL3 is no longer required

11 17/12/2007 PPD Christmas Lectures 2007 New Monitoring and Configuration Tools Nagios –Actively monitors hosts and services and sends alerts when things go wrong Cfengine –Configuration management tool –Change central config and all nodes automagically pick up the changes Pakiti –Monitors the patch status of nodes –May be pushed out to other linux boxes in the Department

12 17/12/2007 PPD Christmas Lectures 2007 Tier 3 Integration with the Tier 2 We’ve taken a number of steps to integrate the Tier 3 with the Tier 2 –User account databases have been merged across both services –Disk mounts are shared between the services Local home, software and scratch areas are mounted on the Grid Workers The grid software and data areas are available on the frontends –SL4 Front end are pbs clients for the Grid batch system Allows direct submission of jobs from the SL4 front ends to the grid batch workers

Tier 3 Services

14 17/12/2007 PPD Christmas Lectures 2007 RALPP and SouthGrid VOs For projects without a VO infrastructure –Either just getting going or too small or short to warrant setting one up RALPP VO is purely in the department SouthGrid will be supported at other SouthGrid sites if you need access to more resources

15 17/12/2007 PPD Christmas Lectures Types of Disk Space Home Areas: –RAID Disks, backed up to tape daily and mirrored to backup server every 12 hours Experiment Areas: /opt/ppd/ –RAID Disks, mirrored to backup server every day, not backed up Data Areas: /pnfs/pp.rl.ac.uk/data/ –dCache, multiple RAID servers, single copy not backed up.

16 17/12/2007 PPD Christmas Lectures Types of Disk Space NFS Scratch: /opt/ppd/scratch –RAID for speed/aggregation, no mirroring or backup Local Scratch: –Spare local disk on front ends and batch workers /scratch on frontends – no guarantees $WORKDIR on batch workers, cleaned up after job finishes

17 17/12/2007 PPD Christmas Lectures 2007 Data Storage Will provide large data areas to Projects/Experiments via the dCache storage element –Write access Grid tools –Read access Grid tools dcap Xrootd Experiment areas for the Main supported VOs RALPP catchall for other users/projects

18 17/12/2007 PPD Christmas Lectures 2007 Home areas quotas All home areas com with a quota quota -v Default quota is small, just 20MB –Many accounts never exceed this –Can exceed this up to 2GB for up to 90 days –If you need it increased, just ask

19 17/12/2007 PPD Christmas Lectures 2007 Batch Submission for SL4 frontends Batch cluster uses PBS like CSF The local access queue is prod The default memory and walltime limits are low, if you need more, specify it: qsub –q prod \ –l mem=1024mb \ –l walltime=24:00:00 \ my-script.sh

20 17/12/2007 PPD Christmas Lectures 2007 Printing Printing migrated from using the BTID print server to the new Departmental print server All departmental PostScript printers should be available lpr –P my-file.ps

21 17/12/2007 PPD Christmas Lectures 2007 Access from Offsite Currently blocked to all nodes –Either use PPTP to tunnel in or ssh via the RAL Bastion host ( Currently looking at the possibility of running some sort of departmental bastion –Separate account database? –Restricted function? –Ssh keys only? –Gsissh only? –Scp/sftp only?

22 17/12/2007 PPD Christmas Lectures 2007 Future Plans Upgrade UIs –Faster, 64bit nodes More batch and disk capacity Bastion Host Subversion Code Repository? Wiki? ?

23 17/12/2007 PPD Christmas Lectures 2007 Christmas Shutdown (or lack of) All the Tier 2 and Tier 3 services will be running over the Christmas Break Service is “at risk”, I might log in occasionally to check on things, I might even read my every now and again but don’t count on support.

24 17/12/2007 PPD Christmas Lectures 2007 Conclusion Aim to provide scientific computing infrastructure for the department and the wider community Evolving in view of external changes Are we ready for the next year? –You tell us… –…but hopefully we are at least on the way