Presentation is loading. Please wait.

Presentation is loading. Please wait.

INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org Round the regions ROC managers Lyon.

Similar presentations


Presentation on theme: "INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org Round the regions ROC managers Lyon."— Presentation transcript:

1 INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org Round the regions ROC managers Lyon

2 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 2 ROC France Our priority issues for ARM-3: –« ROC On-Duty » vs CIC On-Duty –Support issues ServicesCount VOs Globa l Regio nal VO Server220 RLS/RMC110 RB440 BDII4-- UI>8-- Resources Magnitud e VOs Globa l Regio nal Sites880 Estimated CPUs 1070 0 Job slots2465 0 Storage1,5 To 0 Mass Storage Systems 110

3 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 3 Italian Production Grid – Resources CPU e storage installed (as publicated from GIIS) (**) many WNs are Hyperthreaded

4 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 4 Italian CIC/ROC – Services Resource Brokers: –EGEE/LCG infrastructure  egee-rb-01.cnaf.infn.it ;  grid008g.cnaf.infn.it (DAG enabled) –Italian Production Grid  edt003.cnaf.infn.it –ATLAS VO  egee-rb-02.cnaf.infn.it BDIIs: –EGEE/LCG infrastructure  egee-bdii-01.cnaf.infn.it –Italian Production Grid  ibm140.cnaf.infn.it –ATLAS VO  egee-bdii-02.cnaf.infn.it Replica Location Service for babar, virgo, cdf, planck and other Italian VOs –datatag2.cnaf.infn.it, LDAP SERVER FOR National Vos (bio, inaf, ingv, gridit,theophys, virgo) : –grid-vo.cnaf.infn.it

5 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 5 Italian CIC/ROC – Services MyProxy servers: –testbed013.cnaf.infn.it User Interfaces: –UIs are not Core servives..., anyway you can find a list of italian UIs  http://grid-it.cnaf.infn.it/index.php?userinterface&type=1http://grid-it.cnaf.infn.it/index.php?userinterface&type=1 Monitoring:GridICE server: –EGEE/LCG Production infrastructure  http://gridice2.cnaf.infn.it:50080/gridice/site/site.php http://gridice2.cnaf.infn.it:50080/gridice/site/site.php –Italian Production Infrastructure  http://edt002.cnaf.infn.it:50080/gridice/site/site.php http://edt002.cnaf.infn.it:50080/gridice/site/site.php –Atlas  http://grid014.mi.infn.it:50080/gridice/site/site.php –GILDA  http://alifarm7.ct.infn.it/gridice/site/site.php Voms server: –testbed008.cnaf.infn.it –VOs: infngrid, zeus, cdf, planck, compchem

6 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 6 ROC UKI ServicesCount VOs Globa l Regio nal VO Server1/2 0 RLS/RMC000 RB220 BDII22- UI>8-- Resources Magnitud e VOs Globa l Regio nal Sites19 0 Estimated CPUs 1820 0 Job slots1820 0 Storage25TB 0 Mass Storage Systems 110

7 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 7 UKI activities Now hold biweekly Deployment team meetings & monthly deployment meetings with all sites Training – repeat of LCG system administrators course later this year Support desk (integration with GGUS continues) Contributing to User Support on Duty work DSA 1.6 - just starting Preparation for LCG Service Challenges (SRM deployment) NGS reviewing gLite Products: Imperial College working on Sun Grid Engine information providers Automated porting (Grid Ireland) –Several man months effort (TCD) is non-EGEE funded. Supported by RAL, CERN & INFN. –Ports to Fedora, MacOS and AIX (last two require 64-bit versions of VDT). For 2.3.0 need WMS and RGMA from gLite ported Web pages –GridPP deployment pages (http://www.gridpp.ac.uk/deployment) –Web page data sharing mechanism for security content –Reviewing content of UKI ROC pages UKI testzone Network monitoring tools CIC work: Setting up VOMS for PhenoGrid CIC on Duty EGEE general Improvements to GOC portal (new version this week: http://goc.grid-support.ac.uk/gridsite/gocdb/gocdb2.html )http://goc.grid-support.ac.uk/gridsite/gocdb/gocdb2.html Improving Accounting portal

8 INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org SA1 Status Report D/CH Sven Hermann, Holger Marten Forschungszentrum Karlsruhe

9 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 9 Production Service in D/CH Resource Centres (EGEE partners) Additional Centres (non-EGEE partners) No detailed resources committed but wish to follow mw installation/certification procedures Planned resources D/CH: PM 14 (June ‘04 = 400 CPUs, 64 TB disk, 115 TB Tape) 10 - 15 75 4 0,5 ? 9 S. Plan [Tb] Debian SL 3.0.x OS LCG 2.3 MW 336 1070 30 48 36 (88) 4 CPUs 0,07 62 0,26 0,5 0,5 (?) 0,15 Storage [Tb] RH7.3 & SLC3 WNs yes FZK Need for Debian yes GSI FhG / SCAI Int./ext. networkFhG / ITWM RH7.3 WNs yes DESY CSCS RemarksTape SUSE SL 3.0.x OS LCG 2.3 MW 4 6 3 3 CPUs 0,5 0,03 ? 0,2 Storage [Tb] Atlas, D0, Need for SUSE 2 Uni Wuppertal CMS, D0 4 RWTH Aachen Campus firewall ? HU Berlin CDF, CMS ? EKP Uni Karlsruhe RemarksS. Plan [Tb]

10 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 10 D/CH resource allocation policy FZKDESYGSI FhG SCAI FhG ITWM HEPyes on demand Bio Medicineyes Earth Scienceyes Computational Chemistry Astrophysics yes (MAGIC) yes Others Synchrotron X-FEL 19 supported VOs, number unchanged since QR2 about 97% of resources for HEP

11 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 11 Production & other services in D/CH Production Service –FZK  4.2 Million user hours, 1.5 Million Jobs in 2004  65% of resources used by non-LHC (BaBar, CDF, Dzero, Compass)  35% of resources used for contributions to LCG DCs (partly via LCG) –DESY  running Monte Carlo simulations for ZEUS in collaboration with 22 sites, 1800 CPUs, 100 Mio events simulated  H1 preparing MC production as well –FhG/SCAI  users run applications for BioMed & ESR Pre-Production Service At FZK –LCG 2.3.0 under SLC3; moving to LCG 2.4.0, then gLite Rotating MW Installation Support In D/CH (last year)

12 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 12 SA1 Support/Operation ROC Operations Support D/CH –Handle tickets created in CERN/Savannah –Planned: Web-Service like GGUS –Now:  Create contact Mail “ROC-on-duty” with auto-forward to support group  Support group changes every two weeks EGEE 2nd Level Support likely with ROC on duty operation support General Problems: -Rollout gLite: When? Where (testbed, pre-prod)? 2 Versions? -Large clusters: need 2 versions (of OS+MW) at same time, migration to new versions difficult -Monitoring web pages not very stable (problem tracking difficult) -Information on available storage inconsistent (GIIS) FZK11/04/2005 - 24/4/200515/1654/55 DESY28/03/2005 - 10/4/200513/1452/53 GSI14/03/2005 - 27/3/200511/1250/51 FhG28/02/2005 - 13/3/20059/1048/49 FZK14/02/2005 - 27/2/20057/846/47 On Duty Site / Contact DateCalendar Week Project Week

13 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 13 NE Resources (1) Netherlands (NLGrid) – two sites – 366 WN CPUs – 2 disk SEs: 7 TB, mass storage SE: 20TB tape – 2 RBs, RLS and VO LDAP servers for NL-Grid and EGEE – VO support  National: NCF, PVIER, ASTRON, Nadc, VLe, Asci  EGEE: ALICE, ATLAS, LHCb, CMS, Dzero, ESR, Magic – Cluster with 544 Xeon™ CPUs may become CE (with Debian OS) later this year Belgium (BEGrid) – 6 sites running LCG2 software, not integrated in EGEE yet

14 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 14 NE Resources (2) Sweden (SNIC/SWEGRID) Three sites HPC2N  100 CPU cluster  LCG2.3, SLC CE, Debian WN  SE 10 TB  Worker nodes are shared between EGEE and SweGrid (ARC) – PDC  Installation in preparation of a 100 CPU system  Eventually new 884 CPU Xeon EM64T cluster (in stable gLite time) – NSC  32 CPU cluster on pre-production testbed SLC all over LCG23 – Other stuff ( GGAS (Grid bank), SWEGRID VO Management tool) – Finish RC earliest April – VO support  National: Atlas, NorduGrid (used for individual scientists using SweGrid)  To be transferred to EGEE

15 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 15 NE Issues Many answers can be found (for LCG) in the LCG-ROLLOUT archives – can we consolidate? Accounting: we need filters to extract VO specific information CIC on Duty – communication sometimes difficult – No response to questions/remarks Local support not yet centralized

16 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 16 ROC SWE (Portugal and Spain) ServicesCount VOs Globa l Regio nal VO Server220 RLS/RMC220 RB13 0 BDII13 - UI+16 - Resources Magnitud e VOs Globa l Regio nal Sites13 0 Estimated CPUs 660 0 Job slots660 0 Storage2 Tb 0 Mass Storage Systems 220

17 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 17 ROC SEE Our priority issues for ARM-3: –Clarify procedures, such as SLAs, VOs, negotiation etc –Clarification of roles and responsibilities –Support –Reliability / fail-over capabilities of services ServicesCount VO Server1 RLS/RMC0 RB2 BDII2 UI>5 Monitor GridIce, SFT ResourcesMagnitude Sites10 Est. CPUs169 Storage4,8 Tb Mass Storage1 SAN VOs HEP, BioMed, ES, regional SEE-VO

18 INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org CE ROC Deployment – CYFRONET Krakow Monitoring & Operational Support – PCSN Poznan User Support – ICM Warsaw Andrzej Ozieblo ACC CYFRONET AGH

19 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 19 Deployment Working Group Current state –8 RCs operational  IA64 LCG-2-based (PSNC Poznan) waiting for certification  3 new pending RCs (Slovakia, Austria, Hungary) –Certification testbed – for T&V software added locally –pre-production service deployed at CYFRONET (gLite)

20 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 20 DWG Tasks Release customization (Certification Testbed) –prepare customized release according to VO/regional needs  Currently three software packes ready: glogin, ocm-g, g-pm application monitoring tools  Support for MPICH (soon)  Application from II-SAS – flood simulation (soon)  APT-compliant repository ready  Installation using YAIM –VOCE ??? Coordination of activity: on-demand VRVS meetings + RC admins mailing list

21 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 21 Monitoring and operational support Monitoring and managing tasks for CE ROC in Savannah CERN. ROC cluster tests: –Very draft version: http://ras.crossgrid.man.poznan.pl/egee/monitoring/ –3 kinds of monitoring:  mapcenter – testing service responces  testbed status – monitoring informations from site MDS – job/CPU statistic  site tests – tailored and modified tests from LCG testzone report. Installation of LCG2.3 on IA64 clusters – cooperation with Andreas Unterkircher. –It seems that new installation works (still some problems with R-GMA); waiting for add toTestzone.

22 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 22

23 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 23

24 Enabling Grids for E-sciencE INFSO-RI-508833 EGEE ARM-3, 20050317 24 ROC Russia Our priority issues for ARM-3: –« ROC On-Duty » vs CIC On-Duty –Support issues ServicesCount VOs Globa l Regio nal VO Server101 RLS/RMC101 RB101 BDII101 UI>9- Resources Magnitud e VOs Globa l Regio nal Sites960 Estimated CPUs >255 0 Storage7,8 TB 0 Mass Storage Systems 220


Download ppt "INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org Round the regions ROC managers Lyon."

Similar presentations


Ads by Google