T. Bowcock A.Moreton, M.McCubbin CERN-IT 5/00. 29 May 2000CERN-IT T. Bowcock2 University of Liverpool MAP System COMPASS Grid Summary.

Slides:



Advertisements
Similar presentations
Chapter 3: Introduction to Data Communications and Networking
Advertisements

S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
Fighting Malaria With The Grid. Computing on The Grid The Internet allows users to share information across vast geographical distances. Using similar.
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
23rd April 2002HEPSYSMAN April Oxford University Particle Physics Site Report Pete Gronbech Systems Manager.
UCL HEP Computing Status HEPSYSMAN, RAL,
05/07/00LHCb Computing Model Meeting (CERN) LHCb(UK) Computing Status Glenn Patrick Prototype (Tier 1) UK national computing centre Bid to Joint.
24-Apr-03UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS SECURITY SOFTWARE MAINTENANCE BACKUP.
VELO T. Bowcock1 Si Status T.Bowcock for S. Biagi, G.Casse, P. Cooke, M. McCubbin, A. Muir, J. Palacios, C. Parkes, U. Parzefal, G. Patel, J. Phillips.
T. Bowcock1 Can CP save life? LHCb and its applications…
5 June 2000Valencia T. Bowcock1 Themis Bowcock Valencia June 00.
T. Bowcock Liverpool December Nov-00T. Bowcock University of Liverpool Status CDF/GRID.
The IP Revolution. Page 2 The IP Revolution IP Revolution Why now? The 3 Pillars of the IP Revolution How IP changes everything.
LHCb Computing Activities in UK Current activities UK GRID activities RICH s/w activities.
GridPP News NeSC opening “Media” dissemination Tier 1/A hardware Web pages Collaboration meetings Nick Brook University of Bristol.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
 Changes to sources of funding for computing in the UK.  Past and present computing resources.  Future plans for computing developments. UK Status &
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
11 Dec 2000F Harris Datagrid Testbed meeting at Milan 1 LHCb ‘use-case’ - distributed MC production
Fundamentals of Networking Discovery 1, Chapter 2 Operating Systems.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Nick Brook Current status Future Collaboration Plans Future UK plans.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
T. Bowcock Liverpool Sept 00. Sept LHCb-GRID T. Bowcock 2 University of Liverpool Successes Issues Improving the system Comments.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Computing for LHCb-Italy Domenico Galli, Umberto Marconi and Vincenzo Vagnoni Genève, January 17, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
MAP Project T. Bowcock, A. Kinvig, I. Last M. McCubbin, A. Moreton C. Parkes, G. Patel University of Liverpool.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
Spending Plans and Schedule Jae Yu July 26, 2002.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
…building the next IT revolution From Web to Grid…
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Grid Glasgow Outline LHC Computing at a Glance Glasgow Starting Point LHC Computing Challenge CPU Intensive Applications Timeline ScotGRID.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
LHCb computing model and the planned exploitation of the GRID Eric van Herwijnen, Frank Harris Monday, 17 July 2000.
UK GridPP Tier-1/A Centre at CLRC
Nuclear Physics Data Management Needs Bruce G. Gibbard
Gridifying the LHCb Monte Carlo production system
LHCb thinking on Regional Centres and Related activities (GRIDs)
Presentation transcript:

T. Bowcock A.Moreton, M.McCubbin CERN-IT 5/00

29 May 2000CERN-IT T. Bowcock2 University of Liverpool MAP System COMPASS Grid Summary

29 May 2000CERN-IT T. Bowcock3 LHCb Experiment –CP violation –Rare B decays –signals of 10 3 to 10 6 Backgrounds –Potentially all collisions/year! 32 2 About BB produced/year

29 May 2000CERN-IT T. Bowcock4 LHCb Experiment Vertex detector

29 May 2000CERN-IT T. Bowcock5 LHCb Experiment Optimize the Detector Study the Backgrounds

29 May 2000CERN-IT T. Bowcock6 Simulation Full GEANT3 simulation –Event takes of order s on a 400MHz PC Put together a simulation facility –Samples of 10 7 to 10 8 / year –Many times more passed through GEANT –Monte Carlo Array Processor –Similar or larger samples –10 9 institute/year Analysis, reprocessing

29 May 2000CERN-IT T. Bowcock7 Philosophy Fixed Purpose (MC): simplicity Low Cost –No Gbit ethernet until price falls –Dont buy top of range processors –No SMP boards 1998/1999 –No tapes Develop architecture with future in mind –Minimum maintenance/development

29 May 2000CERN-IT T. Bowcock8 Using MAP Disposable MC(throwaway!) Cost Write out ntuple/summary information I/O not really limited by architecture Events may be written out Small internal disks

29 May 2000CERN-IT T. Bowcock9 Hardware 300 processors –400MHz PII –128 Mbytes memory –3 Gbytes disk/processor (IDE) –D-Link 100BaseT ethernet +hubs –commercial units custom boxes for packing and cooling –Total 600kChF inc 17.5% VAT 1998/1999 (Funding Jan 99). ITS Including installation and 3-yr next day on-site maintenance.

29 May 2000CERN-IT T. Bowcock10 MAP-OS Linux –Originally RH5.2 (also tested 6.1) –Stripped to minimum On disk 180MBytes! –Will (with FCS) reinstall/upgrade itself –Access/security

29 May 2000CERN-IT T. Bowcock11 View High Gflops/m 2 Old Mainframe Room Power supply (3 phase) 0.1MW max 50kw cooling

29 May 2000CERN-IT T. Bowcock12 Architecture Master External Ethernet MAP Slaves Hub (Switch- 00) Hub (Switch - 00) 100BaseT

29 May 2000CERN-IT T. Bowcock13 Design Features Mother boards/bios –No keyboard etc required on boot! Front panels –All connections except power Access to each PC via trolley on wheels Cheaper than patch panel! Very convenient. –Cooling (room air flow) 30kW required 50kW capacity Power cutoff installed Rack Mount –30/rack, easy to extract

29 May 2000CERN-IT T. Bowcock14 Learned… Prototype Cables –Cheaper ethernet cables seem OK Would have been nice to have –On board power/heat sensing Dont really need power system –Daisy chain in groups of 5 –Transients can be huge!

29 May 2000CERN-IT T. Bowcock15 Bad things happen… Catastrophic power failure –No UPS (original design had one) –4% needed manual intervention but no hardware failure Burn-in & 4 months of operation –1 power supply exploded –4 PCs with mother-board problems –5 HD failures (within 1 week of turn on) –NIC cards fail –Typically 1% nodes may have a problem

29 May 2000CERN-IT T. Bowcock16 Flow Control System MAP-FCS –UDP level (frames) –solve packet-loss problem Bad hubs(D-Link) NIC Realtek clones with high failure rate –Broadcast system 4 Mbytes/s 300 (Master to Slaves) –Point to point on fail –Standard Mode Communication only with master –Control up to 10,000 PCs

29 May 2000CERN-IT T. Bowcock17 Performance Jan/May 00 –15 million GEANT events for optimization –cf 250,000 possible at CERN –DELPHI events 500,000/day Trilinear Gauge Couplings, W-mass systematics –ATLAS, CDF, H1

29 May 2000CERN-IT T. Bowcock18 User Interface to master only –Web/Grid interface –Security Submission script –Job Control File Sequential jobs, files to keep etc Quick and easy to use Statically linked executable Toolkit –Enables assembly/merging of 300 outputs

29 May 2000CERN-IT T. Bowcock19 Search Analysis As a search-engine MAP architecture is ideal –Low search and recovery times –Chemistry Centre for Innovative Catalysis (JIF 00), promises world lead for Liverpool. –Bio-informatics Compute/search farms

29 May 2000CERN-IT T. Bowcock20 Extending MAP Wish to store events –Part of our mindset (reevaluate?) With existing system –Build an analysis and storage system –Add on disk servers

29 May 2000CERN-IT T. Bowcock21 COMPASS

29 May 2000CERN-IT T. Bowcock22 COMPASS-99 DELL ITS

29 May 2000CERN-IT T. Bowcock23 COMPASS-00 3Tbytes –On top of 1TByte MAP internal Rack Mounted Prototype of 40Tbyte system

29 May 2000CERN-IT T. Bowcock24 COMPASS Low cost(25KCHf/Tbyte inc 17.5%VAT) –SCSI disks(10 50GByte) –Dual Redundant Power Supplies –No RAID backplane –No hotswap –750 MHz processors + 512MBytes memory –Linux –Act as MAP masters

29 May 2000CERN-IT T. Bowcock25 COMPASS Have 3Tbytes of store for R&D on GRID and exploitation of MAP MAP & COMPASS are complementary… Originally requested 40TBytes of store –For H1, BaBar, ATLAS, DELPHI

29 May 2000CERN-IT T. Bowcock26 MAP &COMPASS DST or processed data stored –From MAP Reprocessed/analysed locally –COMPASS Limit data movement off site –COMPASS farm in own right –Powerful analysis engines –Access from remote sites –Designed to, in parallel, analyse very large data sets (Data split nodes – June 00)

29 May 2000CERN-IT T. Bowcock27 Data Transfer Data transfer to/from –Liverpool-CERN/RAL –Liverpool-SLAC/FNAL High Speed link may be a waste of money –3MCHF for 2MBs line! –Quality of service –Probably not true in long term t Transfer disks

29 May 2000CERN-IT T. Bowcock28 MAP-2001 Extension of existing architecture –Vast underestimate of amount of MC required –Extend to 1000 PCs MHz PIII with 72Gbyte disks 128MBytes memory Switched network (&higher quality!) Better NICs/(onboard?)

29 May 2000CERN-IT T. Bowcock29 MAP-2001 Companies more willing to discuss COTS type architecture –Many selling BEOWULF systems –Even IBM! –ITS will provide a turnkey system including our version of MAP control

29 May 2000CERN-IT T. Bowcock30 MAP-2001 Capability –Standard MAP mode –DST transfer –Search Engine –Interprocess communication –Large Internal Store Minimize network traffic Reprocessing

29 May 2000CERN-IT T. Bowcock31 MAP-2001 Increase power by factor of 5 Aim for 1.5M LHCb events/day –Non-volatile 1 Tbyte/day –50Days internal store Use for reprocessing data Disk size will increase by calendar 2001 Multi-user and projects

29 May 2000CERN-IT T. Bowcock32 Issues Authentication and Security Quality of Service Resource Allocation

29 May 2000CERN-IT T. Bowcock33 Grid Adding Globus (June 2000) Access from CERN & –Cambridge University, JMU, Liverpool, RAL Remote submission

29 May 2000CERN-IT T. Bowcock34 Grid 2005 Tier 1 T Tier 0 (CERN) ?????? T2

29 May 2000CERN-IT T. Bowcock35 Grid-LHCb Aim to use MAP as an LHCb testbed –MC production –Data access –Analysis –UK and CERN sites –Interaction with RAL

29 May 2000CERN-IT T. Bowcock36 Health Grid Virtual Population Laboratory –Co-proposed by Liverpool for a world scale met office for disease prediction in collaboration with WHO –Analysis power based on MAP 5000 PC system

29 May 2000CERN-IT T. Bowcock37 Health Grid Community Health Surveillance –WAP, local data bases Information –statistics, Analysis –MAP like centres for Health Policy WHO Med Centre

29 May 2000CERN-IT T. Bowcock38 Comments High Power MC systems vital for HEP –Do we have/plan enough for LHC? Cost and Techniques of Storage –Small groups cant afford/want HSM –Is tape obsolete? Problems for institutes not the same as for Tier 0/1 centres

29 May 2000CERN-IT T. Bowcock39 Summary MAP fulfils its design goals (works!) –MAP-FCS control up to 10,000PCs Minimum manpower 0.5FTE to date –Maintenance and development COTS architecture a success –Low cost has its ups and downs! –MAP available off the shelf for HEP-MC Low cost high density storage server farm in prototype Grid enabled –Access from CERN – and UK HEP institutes soon MAP-2001 –Test a 1000PC farm for LHC