Breaking the frontiers of the Grid R. Graciani EGI TF 2012.

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Advertisements

High Performance Computing Course Notes Grid Computing.
DIRAC Distributed Computing Services
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
1 Managing distributed computing resources with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille September 2011, NEC’11, Varna.
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
BESIII distributed computing and VMDIRAC
Grid Initiatives for e-Science virtual communities in Europe and Latin America DIRAC TEAM CPPM – CNRS DIRAC Grid Middleware.
YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
Enabling Grids for E-sciencE ENEA and the EGEE project gLite and interoperability Andrea Santoro, Carlo Sciò Enea Frascati, 22 November.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
The GridPP DIRAC project DIRAC for non-LHC communities.
GRID Security & DIRAC A. Casajus R. Graciani A. Tsaregorodtsev.
DIRAC Pilot Jobs A. Casajus, R. Graciani, A. Tsaregorodtsev for the LHCb DIRAC team Pilot Framework and the DIRAC WMS DIRAC Workload Management System.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
VLDATA Common solution for the (very-)large data challenge EINFRA-1, focus on topics (4) & (5)
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
EGI Technical Forum Amsterdam, 16 September 2010 Sylvain Reynaud.
Enabling Grids for E-sciencE Experience Supporting the Integration of LHC Experiments Computing Systems with the LCG Middleware Simone.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
The GridPP DIRAC project DIRAC for non-LHC communities.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
LHCb/DIRAC week A.Tsaregorodtsev, CPPM 7 April 2011.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
DIRAC in LHCb and beyond Philippe Charpentier (LHCb-CERN) Slides courtesy of A.Tsaregorodtsev BigPanda Workshop, October Click to edit Master title.
DIRAC as a Service R. Graciani 1, V. Méndez 2, T. Fifield 3, A. Tsaregordtsev 4 1 University of Barcelona 2 University Autónoma of Barcelona 3 University.
DIRAC Distributed Computing Services A. Tsaregorodtsev, CPPM-IN2P3-CNRS FCPPL Meeting, 29 March 2013, Nanjing.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
1 Building application portals with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 27 April 2010, Journée LuminyGrid, Marseille.
Multi-community e-Science service connecting grids & clouds R. Graciani 1, V. Méndez 2, T. Fifield 3, A. Tsaregordtsev 4 1 University of Barcelona 2 University.
Distributed Computing Framework A. Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille EGI Webinar, 7 June 2016.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Bob Jones EGEE Technical Director
Overview of the Belle II computing
DIRAC services.
Grid Deployment Board meeting, 8 November 2006, CERN
WLCG Collaboration Workshop;
Monitoring of the infrastructure from the VO perspective
gLite The EGEE Middleware Distribution
The LHCb Computing Data Challenge DC06
Presentation transcript:

Breaking the frontiers of the Grid R. Graciani EGI TF 2012

Outline  Introduction  The DIRAC Interware  DIRAC as a Framework/Service  Summary Prague, Sep 20th 20122EGI TF 2012

Distributed Computing  Distributed Computing is about splitting up a Computing Task into parts that run on multiple computers communicating over a network.  The main goal of a Distributed Computing system is to connect users to resources in a transparent, open, and scalable way. Prague, Sep 20th 20123EGI TF 2012

Grid proven to work  Discovery of the Higgs boson (July 2012).  A collaborative work of thousands of physicists, sharing resources on hundreds of resource centers.  Results available shortly after the experiments collected the data.  Key role of Grid on this success. Prague, Sep 20th 20124EGI TF 2012

Known issues  End user:  Complicated interfaces for non-experts  Large communities:  Many heterogeneous resources  Intra-community policies  Other communities:  Lack of support for collective work  Resource Providers:  Large: avoid specific configurations and services  Small: reduce overhead of a full grid installation Prague, Sep 20th 20125EGI TF 2012

The DIRAC Interware Prague, Sep 20th 20126EGI TF 2012

DIRAC Solution Prague, Sep 20th 2012EGI TF  LHC experiments developed their own middleware to address the above problems  DIRAC was developed originally for the LHCb experiment  DIRAC is providing a complete interware solution with the goal:  Integrate all the heterogeneous computing resources available  Provides both WMS and DMS  Make the grid convenient for the users:  Single interface  Allowing interoperability  Enabling Community policies  Minimize human intervention at sites Grid A Grid B (WLCG) (NDG) Grid A Cloud B

Workload Management Prague, Sep 20th 2012EGI TF  Jobs are submitted to the DIRAC Central Task Queue with credentials of their owner  Pilot Jobs are submitted by specific Directors to computing resources using appropriated credentials  The Pilot Job fetches the user job and the job owner’s proxy  The User Job is executed with its owner’s proxy used to access SE, catalogs, etc

Physicist User EGEE Pilot Director EGEE Grid NDG Pilot Director NDG Grid Cloud VM Director Cloud CREAM Pilot Director CREAM CE Matcher Service Production Manager Prague, Sep 20th 2012EGI TF 20129

Data Management Components Prague, Sep 20th 2012EGI TF  Storage Elements  gLite/EGI Storage Elements (SRM/gsiftp)  DIRAC Storage Elements  More Storage Elements can be included  (F,SF,HT,BBF)TP servers  File Catalogs  LCG File Catalog (LFC)  DIRAC File Catalog  Support for the User Metadata  Support for data provenance  More Catalogs can be included  LHCb has developed several specific catalogs in the same framework

Data Management Prague, Sep 20th 2012EGI TF  Asynchronous data operations  transfers, registration, removal  Two complementary replication mechanisms  Transfer Agent  user data  public network  FTS service  Production data  Private FTS OPN network  Smart pluggable replication strategies

User Interfaces  Unix command line interface  Historically the first one  Suitable for unix addicts  Easy to use in shell scripts  Too many commands  Even more switches  Python API  Originally for DIRAC developers  More users are starting to use it  The most versatile  Build your own DIRAC applications Prague, Sep 20th EGI TF 2012

And Web Portal Prague, Sep 20th 2012EGI TF

Integration with other tools  DIRAC uses a private API for the secure connection of all its components  A prototype RESTful API has been implemented  Tested to integrate with Gateways or mobile applications  Currently working on the production version  Based on OAuth v2 Prague, Sep 20th EGI TF 2012

It is a Framework  Other necessary components are integrated :  Information system  User / Resource management  Accounting  Monitoring  Can be easily extended  Not all use cases will be covered out of the box  Develop new services, agents, portal pages, scripts,…  May be incorporated to Core functionality  DIRAC is now an open source project   Prague, Sep 20th EGI TF 2012

DIRAC as a Framework/Service Prague, Sep 20th EGI TF 2012

Large communities  Dedicated installations  Several large communities are using DIRAC to implement their computing model  LHCb (CERN)  ILC (CERN)  Belle II (KEK, Japan)  BES III (IHEP, China)  CTA  …  Use exiting functionality, develop ad-hoc extensions, and contribute new functionality Prague, Sep 20th EGI TF 2012

LHCb, the main user Prague, Sep 20th EGI TF 2012

ILC (CERN) Prague, Sep 20th EGI TF 2012

Belle II, combines EGI-OSG Prague, Sep 20th EGI TF 2012

CTA, beyond HEP Prague, Sep 20th EGI TF 2012

BES, building their “Grid” Prague, Sep 20th EGI TF 2012

Why?  Make extensive usage of tools for automated manipulation of large numbers of similar jobs  Simulations  Data processing (workflows / pipelines)  Deep interest on integrated DIRAC File Catalog solution  Replica catalog  Metadata catalog  Provenance catalog  Push development of new interfaces  Interfaces to different batch systems  Combination of different resources Prague, Sep 20th EGI TF 2012

Usage by small communities  Several communities can share a single DIRAC installation operated by a service provider  Currently organized at NGI or Grid level  IN2P3  U. Barcelona  GISELA Latin America Grid  DIRAC provided as a service  Little customization on the installation  Emphasis on dissemination and training Prague, Sep 20th EGI TF 2012

France-Grilles Prague, Sep 20th EGI TF 2012

IberGrid Prague, Sep 20th EGI TF 2012

Summary Prague, Sep 20th 2012EGI TF  DIRAC is an open source software project providing a production- ready solution for distributed computing.  Integrates Grids, Clouds and Clusters, different storage solutions, providing your users with a single interface to all of them.  Ten year HEP experience using the Grid is now available to any interested community.  New functionality and integration of new resource types is added following user needs.  Working to provide a RESTful interface to connect DIRAC with your preferred workflow or application portal solution.

Questions? Prague, Sep 20th 2012EGI TF

Scalability test Prague, Sep 20th EGI TF 2012