Presentation is loading. Please wait.

Presentation is loading. Please wait.

David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\

Similar presentations


Presentation on theme: "David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\"— Presentation transcript:

1 David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ _\/ \/_ _\_\_\/\/_/_/_ / /_/\/\_\ \ _/\/\_ /\ /\

2  John, Rob and Steve have made the local system and the grid work extremely well this year  The result of a very large amount of work on the part of the team means we have had an excellent reliability Local grid performance over the last 3 months

3  Replaced MAP2 with HAMMER  Moved to KVM for grid services  216 helpdesk tickets handled  Helped Maths install their new cluster in our server room  Replaced printers  Made scratch self cleaning and introduced quotas for hepstore  Upgraded user area  Added 40TiB to GridPP storage  All DPM on SL5 and tuned for max performance

4 Since new HEP server [August 2009] was installed the load on email, Twiki etc was much smaller Local batch jobs running: peak was ~85 JanDec Spillover for LCG jobs JanDec

5  The DELL nodes became scrap  Mostly sold at £5 each but other given away  Around 10 tons of scrap removed  The air-conditioning is getting toward the end of its rated life: lower load gives overhead  All racks currently air cooled  Officially not allowed to put more into the server room until next university financial year

6  Linked to 60 nodes in CSD computer room  Dual quad-core AMD Opteron 2.3GHz 'Barcelona' CPUs  32GB memory (4GB per core)  Switch from SUSE 10.3 to CentOS 5 to make the ATLAS software stack happy  10 gigabit link to storage in our cluster room  Will be seen as a separate NW grid site  Jobs run there credited to us  Expect to get a “substantial” portion of the nodes for grid work  Slide is from last year: the link is still used very little

7 Hammercloud is the simulation of the grid usage of ATLAS data analysis jobs Obviously now have real jobs as well but we are still one of the leading sites working on optimising Tier2s for ATLAS jobs

8  LHC and T2K continue data taking  Generally expect an increase in local data storage  Please use DPM for data rather than hepstore or scratch  People should always use the grid for running on large datasets if available  Limited local disk and batch facilities  Can not store a full ATLAS skim here...  Conversion to 64-bit processing for LHC nearly complete  Already moves to use only 64-bit SLC5 nodes on the grid

9  Linux desktop upgrades  SL5/64bit on all Linux analysis machines  Rack/network upgrades for Tier2  Possibly beginning to bottleneck on link to SuperJANET  Storage upgrades  We are behind on our disk quota for GridPP: future money to be spent on that

10

11 Now

12  scratch for rapid access and temporary files  Now cleaned: emails before file deletion  hepstore for medium/long term storage of files  Quotas imposed : 64% full presently  dpm where all large experimental dataset must reside  See the twiki page for how to use it: https://hep.ph.liv.ac.uk/twiki/bin/view/ Computing/GridStorageGuide hepstore free scratch free

13  GridPP 4 grant starts in April next year  Pays for 1.8 people [up from 0.8 last time]  Approved as a 2+2 (i.e. only two years initially)  Large cuts to Tier 2 hardware budgets  Sometime soon the metrics used to determine service levels should be published  Some anomalies ▪ Still asking for 80% of the disk to be empty from before LHC data taking...  If we do well more money flows in future tranches

14  A very successful year for the local team  Excellent performance rewarded by GridPP money  Issues are  Future of 32 bit nodes now very limited  The LHC data taking and analysis is stress testing the grid more thoroughly  Need to maintain excellent performance as small or underperforming sites may not be funded

15


Download ppt "David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\"

Similar presentations


Ads by Google