Computing Infrastructure – Minos 2009/12 ● Downtime schedule – 3 rd Thur monthly ● Dcache upgrades ● Condor / Grid status ● Bluearc performance – cpn lock.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
IHEP Site Status Jingyan Shi, Computing Center, IHEP 2015 Spring HEPiX Workshop.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Jun 29, 20101/25 Storage Evaluation on FG, FC, and GPCF Jun 29, 2010 Gabriele Garzoglio Computing Division, Fermilab Overview Introduction Lustre Evaluation:
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
RAL PPD Site Update and other odds and ends Chris Brew.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Block1 Wrapping Your Nugget Around Distributed Processing.
Tape logging- SAM perspective Doug Benjamin (for the CDF Offline data handling group)
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
Workshop on Computing for Neutrino Experiments - Summary April 24, 2009 Lee Lueking, Heidi Schellman NOvA Collaboration Meeting.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
FermiGrid Keith Chadwick. Overall Deployment Summary 5 Racks in FCC:  3 Dell Racks on FCC1 –Can be relocated to FCC2 in FY2009. –Would prefer a location.
1 LPC Data Organization Robert M. Harris Fermilab With contributions from Jon Bakken, Lothar Bauerdick, Ian Fisk, Dan Green, Dave Mason and Chris Tully.
Nov 18, 20091/15 Grid Storage on FermiGrid and GPCF – Activity Discussion Grid Storage on FermiGrid and GPCF Activity Discussion Nov 18, 2009 Gabriele.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
DBS Monitor and DAN CD Projects Report July 9, 2003.
Status: Central Storage Services CD/LSC/CSI/CSG June 26, 2007.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
IGOR MANDRICHENKO 6/16/2014 Database Applications Update.
Patrick Gartung 1 CMS 101 Mar 2007 Introduction to the User Analysis Facility (UAF) Patrick Gartung - Fermilab.
CD FY10 Budget and Tactical Plan Review FY10 Tactical Plans for Scientific Computing Facilities / General Physics Computing Facility (GPCF) Stu Fuess 06-Oct-2009.
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Minos Computing Infrastructure Arthur Kreymer 1 ● Young Minos issues: – /local/stage1/minosgli temporary reprieve – need grid shared area – Retiring.
Computing Infrastructure – Minos 2009/06 ● MySQL and CVS upgrades ● Hardware deployment ● Condor / Grid status ● /minos/data file server ● Parrot status.
Computing Infrastructure Arthur Kreymer 1 ● Power status in FCC (UPS1) ● Bluearc disk purchase – coming soon ● Planned downtimes – none ! ● Minos.
Keep-up processing Cheryl Patrick MINERvA data management meeting April 8-9, /8/14C Patrick, Keep-up.
Minos Computing Infrastructure Arthur Kreymer 1 ● Young Minos issues: – /local/stage1/minosgli temporary reprieve – need grid shared area – Retiring.
Minos Computing Infrastructure Arthur Kreymer 1 ● Grid – Going to SLF 5, doubled capacity in GPFarm ● Bluearc - performance good, expanding.
Computing Infrastructure Arthur Kreymer 1 ● Power status in FCC (UPS1) ● Bluearc disk purchase – still coming soon ● Planned downtimes – none.
Valencia Cluster status Valencia Cluster status —— Gang Qin Nov
Compute and Storage For the Farm at Jlab
Getting the Most out of Scientific Computing Resources
Status of BESIII Distributed Computing
Getting the Most out of Scientific Computing Resources
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Elizabeth Gallas - Oxford ADC Weekly September 13, 2011
OUHEP STATUS Hardware OUHEP0, 2x Athlon 1GHz, 2 GB, 800GB RAID
Bernd Panzer-Steindel, CERN/IT
Update on Plan for KISTI-GSDC
Castor services at the Tier-0
ATLAS Sites Jamboree, CERN January, 2017
NGS Oracle Service.
US CMS Testbed.
EMC DES-1D11 VCE Test Dumps
The LHCb Computing Data Challenge DC06
Presentation transcript:

Computing Infrastructure – Minos 2009/12 ● Downtime schedule – 3 rd Thur monthly ● Dcache upgrades ● Condor / Grid status ● Bluearc performance – cpn lock program ● Parrot status – moving AFS to Bluearc ● Nucomp and GPCF futures ● Servicedesk issues ● Imap password changes Sep 15 ● kcron fun and games after Aug 20 reboots ● Hardware plans

Computing Infrastructure – Minos 2009/12 DOWNTIMES ● Minos servers got fresh kernels Nov 19 – no further kernel updates needed for at least 60 days ● /minos/data2 locked while copies completed Nov ● PNFS failed Dec 6, 7, and 9 – Hardware problem with the stkensrv1 server, fixed Dec 9 ● PNFS scheduled upgrades Thu Dec 17 07:00-12:00

Computing Infrastructure – Minos 2009/12 DCache issues ● Dcache stable – Server raid controller problem fixed Sep 10. ● Raw data pools expanded Sep 20. – capacity 9.6 TB, good forever ? ● Migration to LTO4 tapes still underway - transparent

Computing Infrastructure – Minos 2009/12 Hardware plans ● Minos Cluster replacements are here – 5 hosts ( 8 cores, 8 TB local disk each ) – Deployment is not scheduled yet – Plan needed for our migration ( code building, etc ) ● Oracle servers being ordered, sharing with Minerva/Nova – Replacements for 6 year old Sun/AMD systems ● FY10 plans are on hold – see budget – Replace remaining older servers ( sam02/3, etc ) – Add 30 TB Bluearc to /minos/data – Add 30 TB to Dcache ( shift from Bluearc )

Computing Infrastructure – Minos 2009/12 Condor/Grid status ● Ryan Patterson managing condor, THANKS !!! ● We continue to reach peaks of 2000 running jobs – still want 5000, after fixing Bluearc issues ● Still using wrapper for condor_rm, condor_hold – To prevent SAZ authentication overload – Condor 7.4 upgrade will fix this internally, next year

Computing Infrastructure – Minos 2009/12 Bluearc performance ● /minos/data2 moved to higher performance disks – The copy tooks months, data rates under 20 MB/sec – Data rates are now around 50 MB/sec. – 4 hours of slow reads in 4 months since D0 separation ● LOCKING - /grid/fermiapp/minos/scripts/cpn – Allows 'n' copies, presently 20, polls in a controlled manner – cpn has locked about 610K times – The old cp1 has locked 360K times so far ● Sep/Oct/Nov locks were 56K / 19K / 4K ● Will symlink cp1 to cpn

Computing Infrastructure – Minos 2009/12 Bluearc performance ● Continuing performance tests – Writes of 660 GB, at sustained 100 Mbytes/sec. – 32 Kbyte blocks on /minos/data2 help fragmentation ● See monitoring plots, such as – - current day – - current week – - historical – - scratch ● Older Minos Cluster nodes read data under 10 MB/sec, will retire them

Computing Infrastructure – Minos 2009/12 PARROT deployment on Fermigrid ● PARROT has problems with SAM projects, etc – Severe problems with I/O intensive activity ( like cp1 or cpn ) – Parrot is designed for CPU limited applications ● We rsynced the parrot-accessed files to /grid/fermiapp and /minos/data ● Robert has made the Minos base released relocatable ● Experts should test this, see....

Computing Infrastructure – Minos 2009/12 NUCOMP / GPCF ● Nucomp monthly coordination meeting - Lee Lueking – These continued, plus workshops ● IF login cluster – Similar to Minos Cluster, but shared, primarily Minerva at present ● GPCF – General Physics Computing Facility – FNALU replacement and Grid front end, for general users – Minerva/Nova/Mu2e will be early customers – Budget calls for about 128 cores interactive, 256 cores batch – Spend money in early FY10