1 WP2: Data Management Gavin McCance RAL Middleware Workshop 24 February 2003.

Slides:



Advertisements
Similar presentations
Giggle: A Framework for Constructing Scalable Replica Location Services Ann Chervenak, Ewa Deelman, Ian Foster, Leanne Guy, Wolfgang Hoschekk, Adriana.
Advertisements

The Replica Location Service In wide area computing systems, it is often desirable to create copies (replicas) of data objects. Replication can be used.
1 WP2: Data Management Paul Millar eScience All Hands Meeting September
WP2: Data Management Gavin McCance University of Glasgow November 5, 2001.
WP2: Data Management Gavin McCance University of Glasgow.
EU DataGrid TestBed 2 Component Review Paul Millar (University of Glasgow) (slides based on a presentation by Erwin Laure)
The Quantum Chromodynamics Grid James Perry, Andrew Jackson, Matthew Egbert, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
OptorSim: A Replica Optimisation Simulator for the EU DataGrid W. H. Bell, D. G. Cameron, R. Carvajal, A. P. Millar, C.Nicholson, K. Stockinger, F. Zini.
ATLAS/LHCb GANGA DEVELOPMENT Introduction Requirements Architecture and design Interfacing to the Grid Ganga prototyping A. Soroko (Oxford), K. Harrison.
WP2 and GridPP UK Simulation W. H. Bell University of Glasgow EDG – WP2.
Data Management Expert Panel - WP2. WP2 Overview.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
N° 1 LCG EDG Data Management Catalogs in LCG James Casey LCG Fellow, IT-DB Group, CERN
DataGrid is a project funded by the European Union CHEP 2003 – March 2003 – Title – n° 1 Grid Data Management in Action Experience in Running and.
GGF Toronto Spitfire A Relational DB Service for the Grid Peter Z. Kunszt European DataGrid Data Management CERN Database Group.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
GridPP9 – 5 February 2004 – Data Management DataGrid is a project funded by the European Union GridPP is funded by PPARC WP2+5: Data and Storage Management.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
ILDG5QCDgrid1 QCDgrid status report UKQCD data grid Chris Maynard.
NAREGI WP4 (Data Grid Environment) Hideo Matsuda Osaka University.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Marianne BargiottiBK Workshop – CERN - 6/12/ Bookkeeping Meta Data catalogue: present status Marianne Bargiotti CERN.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
INFNGrid Constanza Project: Status Report A.Domenici, F.Donno, L.Iannone, G.Pucciani, H.Stockinger CNAF, 6 December 2004 WP3-WP5 FIRB meeting.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Tony Doyle & Gavin McCance - University of Glasgow ATLAS MetaData AMI and Spitfire: Starting Point.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Part Four: The LSC DataGrid Part Four: LSC DataGrid A: Data Replication B: What is the LSC DataGrid? C: The LSCDataFind tool.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
Replica Management Services in the European DataGrid Project Work Package 2 European DataGrid.
SkimData and Replica Catalogue Alessandra Forti BaBar Collaboration Meeting November 13 th 2002 skimData based replica catalogue RLS (Replica Location.
Data Management GridPP and EDG Gavin McCance University of Glasgow May 9, 2002
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Database authentication in CORAL and COOL Database authentication in CORAL and COOL Giacomo Govi Giacomo Govi CERN IT/PSS CERN IT/PSS On behalf of the.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
Data Management The European DataGrid Project Team
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
Site Authorization Service Local Resource Authorization Service (VOX Project) Vijay Sekhri Tanya Levshina Fermilab.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Current Globus Developments Jennifer Schopf, ANL.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
WP2: Data Management Gavin McCance University of Glasgow.
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
Gavin McCance University of Glasgow GridPP2 Workshop, UCL
Spitfire Overview Gavin McCance.
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
Gridifying the LHCb Monte Carlo production system
Presentation transcript:

1 WP2: Data Management Gavin McCance RAL Middleware Workshop 24 February 2003

RAL Workshop Feb Outline  WP2 Tasks  Review of TB1 Components  Changes and review of current components  Plans for final year

RAL Workshop Feb WP2 Tasks  Replication Services Keep track of all the files and their copies Copy them about (to order and automatically)  Optimization of replication Give me the ‘best replica’ for my job Simulate the grid to tune the algorithms needed for this  Meta-data Where will the replication stuff keep its meta-data Where will the applications keep their meta-data  Security Authenticate with grid certificates Authorize users appropriately (better than just a grid-mapfile)

RAL Workshop Feb TB1 Replication: Replica Catalogue  Edg-replica-catalogue The repackaging of the much-loved Globus replica catalogue Based on LDAP  LFN -> PFN (1:many) One logical file name mapping to many physical instances of the file With appropriate utility functions, applications might never need to know the PFN. Use the LFN, and the middleware does the mapping for you in the background.

RAL Workshop Feb TB1 Replication: Copying to order  Edg-replica-manager Initially, it was a repackaging of Globus replica manager  Rewritten for TB1+ with better client interfaces Both command-line and C++ copyAndRegister: ‘brings your new file to the grid’ replicateFile: makes a new replica of a file

RAL Workshop Feb TB1 Replication: Copying ~automatically  GDMP: Grid Data Mirroring Package Born in CMS  Implements subscription-based replication Furious Monte Carlo generation Site A Lots of New files Site B “Subscribe me!” Notify: “I’ve got some new files!” “Send me them” GridFTP of new files Replica catalogue New replicas at site B

RAL Workshop Feb Replication Optimization  Most research-oriented task  Early TB1 getBestFile absent  RB matches LFNs against local storage elements Jobs only go where their data already is No clever data movement  OptorSim developed to test replica optimization ideas Data-centric grid simulation Simulates job times as function of replication mechanism and job data access patterns UK JANet and EU GEANT network modelled

RAL Workshop Feb Meta-data storage  Spitfire meta-data storage  Two faces Spitfire browser Spitfire client API  Spitfire-browser allows a client to use web-browser to view the results of canned queries from a database or make canned inserts into the database.  Client uses their grid cert embedded in their web browser to authenticate (and then authorize) to the service. DB Spitfire Browser Netscape Web browser Fill in the web-page form The result comes back to client

RAL Workshop Feb Meta-data storage 2  Spitfire client API  Imagine where you would use ODBC / JDBC in an application To do something with a database from inside your application  That’s where you use this API, except… Accesses DB over WAN Grid security (both authentication and authz) You shouldn’t have to know what the DB backend is NB. The API is not the same as ODBC!

RAL Workshop Feb Security  WP2 task feeding into EDG security group  Server side: Mostly JAVA Proper certificate trust-manager for java server applications (special plug-in for Tomcat) Flexible authorization manager to define whatever authz policies you like upon the server.  Client side Proper JAVA trust-manager for certificate checking Web services GSI-enabled for Java and C++

RAL Workshop Feb Changes: Web services  Most software has been redesigned to use web services  Much of the server-side stuff now written in Java  Retain security: GSI-enabled web-services  Services have been modified to expose an API in WSDL  For client programming, the client API libraries are auto- generated from the WSDL  For command-line, the tools are still there, but now talk to the server using web services.  What the applications user sees should not have changed as a result of adopting web services!

RAL Workshop Feb Changes: Replica Catalogue to RLS  edg-replica-catalogue being phased out  For Replica Location Service (RLS) collaboration with Globus  Local Replica Catalogs (LRCs) on the SEs hold the actual GUID -> PFN mappings [GUID is what used to be LFN]  Replica Location Indices (RLIs) redirect inquiries to LRCs actually having the file  LRCs are configured to send index updates to any number of RLIs  Much more scalable architecture The lookup time for an entry is independent of the number of catalogs. Tested for up to 10 8 entries. The catalog withstands simultaneous user queries of over 1000 queries or inserts per second.

RAL Workshop Feb RLS Demo at SC2002

RAL Workshop Feb Changes: unified interface for replication  Many services, some the same, some new, with a bewildering array of acronyms…  All these services have their own APIs, and are individually accessible on the grid.  From applications point of view, it’s more appropriate is the have a single client facing interface (both programming and command line) that you can use to talk to all these services. Simpler… you only need to read one document ;-) Allows this single client to take care of transactional issues This is the new EDG Replica Manager (ERM) for TB2

RAL Workshop Feb TB2 Replica Manager Components and name changes… ERM: EDG Replica Manager client interface and API Entry point for all clients ROS: Replication Optimization Service Replica selection based on network metrics (WP7) RSH: Replication Storage Handler (what was GDMP) Subscription-based replication RLS: Replica Location Service (replacing replica catalogue) Local Replica Catalog services LRC: Logical to Physical file mappings Replica Location Index services RLI: index on Logical names RMC:Replication Metadata Catalogue Similar to Spitfire with RDBMS backend and specialized schema NEW !

RAL Workshop Feb TB2: Replica Management Services “Reptor” Replica Management Services Optimization Replica Metadata Subscription Client Replica Location File Transfer ERM RMC ROS RSH TB2 Components GridFTP RLS

RAL Workshop Feb LFNs, PFNs, GUIDs  Due to application requirement from LCG, a couple of changes: PFN1, Glasgow PFN2, CERN PFN3, Lyon GUID ASSDF LFN1 LFN2 LFN3 Replica Location Service Replica Meta-data Catalogue

RAL Workshop Feb TB2: RMC and Spitfire  Replication meta-data considered sufficiently ‘specialized’ and vital to the replica management service that it has been split off from Spitfire Now called Replica Metadata Catalogue (RMC) Resolves LFNs to GUIDs Underlying technology is identical to Spitfire Exposed API is different  More tailored for specific things you’d like to do with replication meta- data. Application specific section for application meta-data that is keyed on LFNs or GUIDs.  Spitfire is still available for other meta-data e.g. storing calibration constants, etc.

RAL Workshop Feb Replica Optimization Service: ROS  Provides getAccessCosts( LFN[], CE[], … ) method to RB Allows RB to take into account the distribution of a job’s files when deciding where to run it  Provides listBestFile ( LFN, toSE ) [in the ERM interface] Uses networking bandwidth + storage cost measurements (WP7 and WP5) to determine the best replica to get.  Provides getBestFile ( LFN, toSE, … ) [in the ERM interface] The same, except it actually does this replication, if needed.  For TB2, simple replication algorithms will be deployed initially. More adventurous ones can be added without impacting the interface, since the replication algorithm is internal to the RMS

RAL Workshop Feb Current OptorSim status  OptorSim used to simulate possible algorithms for ROS  Simulation now includes sampled network background (UCL)  Live network simulator GUI Or offline in a compute farm to get useful results..!

RAL Workshop Feb Current OptorSim results  Initial results from simulation show that including network background increases job times by ~10%.  Further study underway…  Study of different replication algorithms and access patterns  Data access pattern has a large effect Further study here  Economic models do well for sequential data access - 6 experiments, 22 sites - predicted available CPUs & storage - realistic file sizes (1GB) and dataset sizes (1TB) - realistic number of jobs (~60 users) - inclusion of background network traffic

RAL Workshop Feb Plans for final year: Meta-data  RMC is now ~fixed in functionailty  Spitfire will evolve a bit more To allow authorized users to hot-deploy their own interfaces onto the service to do something useful. e.g. you as an analysis-group hardware person can ‘invent’ an method call (an interface) to extract some data from an obscure calibration constant table. Spitfire (which sits in front of DB containing these tables) will then expose your newly invented interface so that people can use it by standard web-services remote procedure call  And web-services will write the client stub for you automatically…  Keep working on OGSA (and GGF DAIS standard)

RAL Workshop Feb Plans for final year: RMS  RMS architecture is now defined  Consolidate and concentrate on quality Few new features  Support LCG Software was developed alongside LCG requirements  Work will continue on improving the algorithms used internally by the ROS (replica optimization)  Work towards EGEE…