The High Energy Physics Community Grid Project Inside D-Grid ACAT 07 Torsten Harenberg - University of Wuppertal

Slides:



Advertisements
Similar presentations
Delta Confidential 1 5/29 – 6/6, 2001 SAP R/3 V4.6c PP Module Order Change Management(OCM)
Advertisements

TAB, 03. March 2006 Bruno Hoeft German LHC – WAN present state – future concept Forschungszentrum Karlsruhe GmbH Institute for Scientific Computing P.O.
Max Mustermann Folientitel Veranstaltung Online Steering of HEP Applications Daniel Lorenz University of Siegen Cracow Grid Workshop –
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
11 Application of CSF4 in Avian Flu Grid: Meta-scheduler CSF4. Lab of Grid Computing and Network Security Jilin University, Changchun, China Hongliang.
Storage Services Let the data flow! NorduNet 2008,.fi, 9 April 2008 Jan Meijer.
Grid Checkpoining Architecture Radosław Januszewski CoreGrid Summer School 2007.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Fabric and Storage Management GridPP Fabric and Storage Management GridPP 24/24 May 2001.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006.
B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
GridPP Presentation to PPARC Grid Steering Committee 26 July 2001 Steve Lloyd Tony Doyle John Gordon.
NGS computation services: API's,
Current status of grids: the need for standards Mike Mineter TOE-NeSC, Edinburgh.
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft LCG-POB, , Reinhard Maschuw1 Grid Computing Centre Karlsruhe - GridKa Regional/Tier.
GridKa January 2005 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann 1 Mass Storage at GridKa Forschungszentrum Karlsruhe GmbH.
Service Data Challenge Meeting, Karlsruhe, Dec 2, 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Plans and outlook at GridKa Forschungszentrum.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
INFSO-RI Enabling Grids for E-sciencE Sven Hermann, Clemens Koerdt Forschungszentrum Karlsruhe ROC DECH Report.
Forschungszentrum Karlsruhe Technik und Umwelt Regional Data and Computing Centre Germany (RDCCG) RDCCG – Regional Computing and Data Center Germany software.
The Platform as a Service Model for Networking Eric Keller, Jennifer Rexford Princeton University INM/WREN 2010.
Page 1 October 31, 2000 An Introduction to Large-Scale Software Development Steve Varnau Core HP-UX Operation October 31, 2000.
25 seconds left…...
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Helmut Dres, Institute For Scientific Computing – GDB Meeting Global Grid User Support.
We will resume in: 25 Minutes.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
STEINBUCH CENTRE FOR COMPUTING - SCC KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
E-Science Workshop, Santiago de Chile, 23./ KIT ( Frank Schmitz Forschungszentrum Karlsruhe Institut.
The German HEP Community Grid for the German HEP Community Grid 27-March-2007, ISGC2007, Taipei Agenda: D-Grid in context HEP Community.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
DESY Participation in an External Experiment Joachim Mnich PRC Meeting
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
IST E-infrastructure shared between Europe and Latin America High Energy Physics Applications in EELA Raquel Pezoa Universidad.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft 1 Institute for Scientific Computing in the Forschungszentrum Karlsruhe Overview Rainer Kupsch.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Computing Coordination Aspects for HEP in Germany International ICFA Workshop on HEP Networking, Grid and Digital Divide Issues for Global e-Science nLCG.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Particle Physics in Germany activities and perspectives Bernhard Spaan - TU Dortmund Chair of Komitee für Elementarteilchenphysik (KET) technische universität.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Storage and Data Movement at FNAL D. Petravick CHEP 2003.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
EGEE is a project funded by the European Union under contract IST Issues from current Experience SA1 Feedback to JRA1 A. Pacheco PIC Barcelona.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
The German HEP-Grid initiative for the German HEP Community Grid 13-Feb-2006, CHEP06, Mumbai Agenda: D-Grid in context.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
DGI: The D-Grid Infrastructure
Grid Computing and the National Analysis Facility
Partner: LMU (Atlas), GSI (Alice)
Leigh Grundhoefer Indiana University
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

The High Energy Physics Community Grid Project Inside D-Grid ACAT 07 Torsten Harenberg - University of Wuppertal

2/27 D-Grid organisational structure

3/27 technical infrastructure Nutzer User API D-Grid resources Grid services Core services Distributed data services D-Grid Services Communities Daten/ Software Distributed computing resources network Security and VO management I/O GAT API Scheduling und Workflow Management Portal (GridSphere based) UNICORE Accounting und Billing Data management Globus Toolkit V4 LCG/gLite Monitoring

4/ EDGEGEE LCG R&DWLCG Ramp-up... EGEE 2 HEP CG Okt. HI run Mar-Sep pp run today EGEE 3 ? GridKa / GGUS DGI HEP Grid effords since 2001 DGI 2 D-Grid Initiative ???

5/27 LHC Groups in Deutschland Alice: Darmstadt, Frankfurt, Heidelberg, Münster ATLAS: Berlin, Bonn, Dortmund, Dresden, Freiburg, Gießen, Heidelberg, Mainz, Mannheim, München, Siegen, Wuppertal CMS: Aachen, Hamburg, Karlsruhe LHCb: Heidelberg, Dortmund

6/27 German HEP institutes participating in WLCG WLCG: Karlsruhe (GridKa & Uni), DESY, GSI, München, Aachen, Wuppertal, Münster, Dortmund, Freiburg

7/27 HEP CG participants: Participants: Uni Dortmund, TU Dresden, LMU München, Uni Siegen, Uni Wuppertal, DESY (Hamburg & Zeuthen), GSI Associated partners: Uni Mainz, HU Berlin, MPI f. Physik München, LRZ München, Uni Karlsruhe, MPI Heidelberg, RZ Garching, John von Neumann Institut für Computing, FZ Karlsruhe, Uni Freiburg, Konrad-Zuse-Zentrum Berlin

8/27 HEP Community Grid WP 1: Data management (dCache) WP 2: Job Monitoring and user support WP 3:distributed data analysis (ganga) ==> Joint venture between physics and computer science

9/27 WP 1: Data management coordination: Patrick Fuhrmann An extensible metadata catalogue for semantical data access: Central service for gauge theory DESY, Humboldt Uni, NIC, ZIB A scaleable storage element: Using dCache on multi-scale installations. DESY, Uni Dortmund E5, FZK, Uni Freiburg Optimized job scheduling in data intensive applications: Data and CPU Co-scheduling Uni Dortmund CEI & E5

10/27 WP 1: Highlights Establishing a metadata catalogue for the gauge theory Production service of a metadata catalogue with > documents. Tools to be used in conjunction with LCG data grid Well established in international collaboration Advancements in data management with new functionality dCache could become quasi standard in WLCG Good documentation and automatic installation procedure helps to provide useability for small Tier-3 installations up to Tier-1 sites. High troughput for large data streams, optimization on quality and load of disk storage systems, giving high performant access to tape systems

11/27 dCache based scaleable storage element dCache project well established New since HEP CG: Professional product management, i.e. code versioning, packaging, user support and test suits. - single host - ~ 10 TeraBytes - Zero Maintenance - thousands of pools - >> PB Disk Storage - >> 100 File transfers/ sec - < 2 FTEs dCache.ORG

12/27 dCache: principle P Backend Tape Storage Streaming Data (gsi)FTP http(g) Posix I/O xRoot dCap Storage Control SRM EIS protocol Engines dCache Controller Managed Disk Storage HSM Adapter dCache.ORG Information Prot.

13/27 dCache: connection to the Grid world Storage Element Firewall IN - SITE Compute Element Information System FTS Channels gsiFtp SRM Storage Resource Manager Protocol File Transfer Service dCap/rfio/root OUT - SITE

14/27 dCache: achieved goals Development of the xRoot protocol for distributed analysis Small sites: automatic installation and configuration (dCache in 10mins) Large sites (> 1 Petabyte): Partitioning of large systems. Transfer optimization from / to tape systems Automatic file replication (freely configurable)

15/27 dCache: Outlook Current usage 7 Tier I centres with up to 900 Tbytes on disk (pre center) plus tape system. (Karlsruhe, Lyon, RAL, Amsterdam, FermiLab, Brookhaven, Nordu Grid) ~ 30 Tier II centres, including all US CMS in USA, planned for US ATLAS. Planned usage dCache is going to be included in the Virtual Data Toolkit (VDT) of the Open Science Grid: proposed storage element in the USA. Planned US Tier I will break the 2 PB boundary end of the year.

16/27 HEP Community Grid WP 1: Data management (dCache) WP 2: Job Monitoring and user support WP 3:distributed data analysis (ganga) ==> Joint venture between physics and computer science

17/27 WP 2: job monitoring and user support co-ordination: Peter Mättig (Wuppertal) Job monitoring- and resource usage visualizer TU Dresden Expert system classifying job failures: Uni Wuppertal, FZK, FH Köln, FH Niederrhein Job online steering: Uni Siegen

18/27 Job monitoring- and resource usage visualizer

19/27 Integration into GridSphere

20/27 Job Execution Monitor in LCG Motivation 1000s of jobs each day in LCG Job status unknown while running Manual error detection: slow and difficult GridICE,...: service/hardware based monitoring Conclusion Monitor job while running JEM Automatical error detection needed expert system

21/27 gLite/LCG Workernode Pre-execution test Script monitoring Information exchange: R-GMA Visualization: e.g. GridSphere Bash Python Experten system for classification Integration into ATLAS Integration into GGUS post D-Grid I:... ? JEM: Job Execution Monitor

22/27 JEM - status Monitoring part ready for use Integration into GANGA (ATLAS/LHCb distributed analysis tool) ongoing Connection to GGUS planned

23/27 HEP Community Grid WP 1: Data management (dCache) WP 2: Job Monitoring and user support WP 3:distributed data analysis (ganga) ==> Joint venture between physics and computer science

24/27 WP 3: distributed data management Co-ordination: Peter Malzacher (GSI Darmstadt) GANGA: distributed ATLAS and LHCb Ganga is an easy-to-use frontend for job definition and management Python, IPython or GUI interface Analysis jobs are automatically splitted into subjobs which are sent to multiple sites in the Grid Data management for in- and output. Distributed output is collected. Allows simple switching between testing on a local batch system and large-scale data processing on distributed resources (Grid) Developed in the context of ATLAS and LHCb Implemented in Python

25/27 GANGA schema Storage queues manager outputs catalog query submit files jobs data file splitting myAna.C merging final analysis

26/27 PROOF schema catalog Storage scheduler query MASTER PROOF query: data file list, myAna.C files final outputs (merged) feedbacks

27/27 DESY, Dortmund Dresden, Freiburg, GSI, München, Siegen, Wuppertal Dortmund, Dresden, Siegen, Wuppertal, ZIB, FH Köln, FH Niederrhein Physics DepartmentsComputer Sciences D-GRID: Germanys contribution to HEP computing: dCache, Monitoring, distributed analysis Effort will continue, 2008: Start of LHC data taking challenge for GRID Concept ==> new tools and developments needed HEPCG: summary