IN2P3Creating Federated Data Stores for the LHC Summary Andrew Hanushevsky SLAC National Accelerator Laboratory September 13-14, 2012 IN2P3,

Slides:



Advertisements
Similar presentations
PRODUCT FOCUS 4/14/14 – 4/25/14 INTRODUCTION Our Product Focus for the next two weeks is Microsoft Office 365. Office 365 is Microsoft’s most successful.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
U.S. Department of the Interior U.S. Geological Survey CDI Webinar Series 2013 Mobile Framework Lorna Schmid, AEI Tim Kern, Fort Collins Science Center.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Google AppEngine. Google App Engine enables you to build and host web apps on the same systems that power Google applications. App Engine offers fast.
Efi.uchicago.edu ci.uchicago.edu FAX status report Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS Computing Integration.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Efi.uchicago.edu ci.uchicago.edu FAX status developments performance future Rob Gardner Yang Wei Andrew Hanushevsky Ilija Vukotic.
MW Readiness WG Update Andrea Manzi Maria Dimou Lionel Cons 10/12/2014.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Storage Federations and FAX (the ATLAS Federation) Wahid Bhimji University of Edinburgh.
Future home directories at CERN
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Julia Andreeva on behalf of the MND section MND review.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
Efi.uchicago.edu ci.uchicago.edu Storage federations, caches & WMS Rob Gardner Computation and Enrico Fermi Institutes University of Chicago BigPanDA Workshop.
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Introduce Caching Technologies using Xrootd Wei Yang 10/28/14ATLAS TIM 2014 Univ. Chicago1.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Ian Bird LCG Project Leader Status of EGEE  EGI transition WLCG LHCC Referees’ meeting 21 st September 2009.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Resource Provisioning EGI_DS WP3 consolidation workshop, CERN Fotis Karayannis, GRNET.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
XRootD Monitoring Report A.Beche D.Giordano. Outlines  Talk 1: XRootD Monitoring Dashboard  Context  Dataflow and deployment model  Database: storage.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Any Data, Anytime, Anywhere Dan Bradley representing the AAA Team At OSG All Hands Meeting March 2013, Indianapolis.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Storage Accounting John Gordon, STFC OMB August 2013.
SLACFederated Storage Workshop Summary Andrew Hanushevsky SLAC National Accelerator Laboratory April 10-11, 2014 SLAC.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
DPM in FAX (ATLAS Federation) Wahid Bhimji University of Edinburgh As well as others in the UK, IT and Elsewhere.
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
Federating Data in the ALICE Experiment
Ricardo Rocha ( on behalf of the DPM team )
Vincenzo Spinoso EGI.eu/INFN
Introduction to Data Management in EGI
PanDA in a Federated Environment
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
EGI UMD Storage Software Repository (Mostly former EMI Software)
Ákos Frohner EGEE'08 September 2008
Logsign All-In-One Security Information and Event Management (SIEM) Solution Built on Azure Improves Security & Business Continuity MICROSOFT AZURE APP.
New Types of Accounting Beyond CPU
SharePoint Online Authentication Patterns
Microsoft Virtual Academy
User Accounting Integration Spreading the Net.
Presentation transcript:

IN2P3Creating Federated Data Stores for the LHC Summary Andrew Hanushevsky SLAC National Accelerator Laboratory September 13-14, 2012 IN2P3, Lyon, France

IN2P3Creating Federated Data Stores for the LHC FAX (ATLAS)  Increase analysis opportunities (WAN, failover, etc)  Ensures only “good” sites join the federation  Adopting regional federation topology  Vector clients to the closest currently available data source  Looking at WAN access vs local caching  Caching seems usually better go but hit-rate and storage issues  LFC look-up is the major stumbling block  Will this be the last time such DM will be developed?  Type of access in user jobs is also a challenge  Uses a rather detailed site certification process  Goal is >90% of available data by the end of the year 2

IN2P3Creating Federated Data Stores for the LHC AAA (CMS)  Increase analysis opportunities (WAN, failover, etc)  Ensures only “good” sites join the federation  Covered the US sites and now working on worldwide sites  WAN access is supportable and works well  Caching seems usually better go but hit-rate and storage issues  Next years tasks include  Hardening  Public cloud usage (depends on pricing)  Data aware job management.  Caching proxy  Client changes take a long time! 3

IN2P3Creating Federated Data Stores for the LHC Panda  Stage in data from the federation as fallback  Could be used as a tape avoidance system  Direct access later (1 st for input, later for output data)  Jobs shipped to available sites with lowest data access cost 4

IN2P3Creating Federated Data Stores for the LHC Need to optimize data access for high latency  TTreeCache is key here (chachinh, async pre-reads, etc) Caching also helps reduce WAN load Monitoring is crucial to track WAN access  Identify badly behaving applications 5

IN2P3Creating Federated Data Stores for the LHC New xroot client available in September EOS being extended for federation & monitoring  Pretty much done, minor additions needed  LFC translation is ready to go for ATLAS DPM (rewritten)  Multi-VO data access that can federate using xroot  Implemented as plug-ins to basic xrootd front-end  Waiting for the final 3.2.x version of xrootd going into EPEL  Will then be available in EMI repository dCache is adding N2N plugins equivalent to the xrootd ones. 6

IN2P3Creating Federated Data Stores for the LHC UCSD Monitoring is in a mature state  ActiveMQ feed almost done (needed for world-wide view) Monitoring useful for measuring data popularity  Rich set of data for mining to see n-order effects  Countless ways to render the data to gain usage insights Information is being rolled into dashboards There is 3-5 months more work to what we need  But how long before people deploy the new stuff? Seems like this will continue to be very active  At least for the next year We need to start thinking about monitoring for multi- VO sites. 7

IN2P3Creating Federated Data Stores for the LHC Federation on cloud storage  Example single data point: 11-16MB/sec/stream on EC2.  Storage: 1TB $ /month  Data flow in cloud 2-3x better than reading from outside There is no real cloud standard  This makes moving from cloud to cloud difficult  EC2 is the biggest player and most popular  Google & Microsoft are showing promise as competitors Using federated storage (in & out) is key  Leaving data in cloud is expensive 8

IN2P3Creating Federated Data Stores for the LHC Publicize the concept of federated storage  Working group to explore concepts (ends at end of year) Separate “federation” from the apps that use it  Allows better exploration of fail-over, self-healing, caching... Biggest issue is the historical lfn->sfn translation  Complicates creating an efficient global name space Federated storage is rapidly progressing  Need still to understand security & monitoring issues Working group still active 9

IN2P3Creating Federated Data Stores for the LHC In year 1 or of 3 year project  Provide storage infrastructure for sharing data Very diverse group of people  Much like a multi-VO system on steriods Decided to use iRODS  Provides federation glued via a database (similar to LFC)  Works well within the confines of itself but has scalability issues.  Does not seem to integrate very naturally with other systems. Looking for other systems (Xrootd, HTTP) to externalize access 10

IN2P3Creating Federated Data Stores for the LHC dCache based storage federation with caching  Data caching is significant activity for worker-node jobs  Competing I/O – caching vs jobs  Cache should be fast for random I/O but managed store need not  Cache also solves a number of operational issues.  Have a rule-of-thumb sizing caches for WLCG sites: 100TB cache per 1PB store. Looking at new ways of federating the caches  Either xrootd or ARC based HTTP Lesson highlights  Have product developers on your own staff  Availability is an upper bound for user’s happiness  One system for all is the (unobtainable) holy grail 11

IN2P3Creating Federated Data Stores for the LHC EMI funded project revolving around DPM  Current system based on Apache + lcgdm_day + dmlite  Plus arbitrary plugins  Development is ongoing to handle edge cases  Endpoint changes (e.g. life, content) Project work to align xroot & http approaches  http plug-in for xrootd signed off  Other possibilities being explored 12

IN2P3Creating Federated Data Stores for the LHC EMI funded project revolving around DPM  Current system based on Apache + lcgdm_day + dmlite  Plus arbitrary plugins  Development is ongoing to handle edge cacses  Endpoint changes (e.g. life, content) Project work to align xroot & http approaches  http plug-in for xrootd signed off  Other possibilities being explored 13

IN2P3Creating Federated Data Stores for the LHC Driving forces for federation  Create more opportunities for data access.  This seems to be strong enough to foster several efforts. Outline broad technical solutions  We outlined several monitoring solutions for cohesiveness  Protocol alignment to allow more inter-play Establish framework for technical co-operation  We have this meeting and WLCG framework  As witnessed by the protocol alignment project Revisit our definition….. 14

IN2P3Creating Federated Data Stores for the LHC Collection of disparate storage resources managed by co-operating but independent administrative domains transparently accessible via a common name space. Maybe we don’t change the definition, but differentiate the things unique to the work discussed here:  Single protocols? Maybe not  From any entry point, access to all the data in the system.  Direct end-user access to files from source? 15

IN2P3Creating Federated Data Stores for the LHC So, should we have another meeting?  If yes, Jeff Templon offered to host at NIKHEF Half-day virtual pre-GDB meeting in April. Dates (green preferred): MTWRFSS November 2013

IN2P3Creating Federated Data Stores for the LHC IN2P3  For hosting this meeting Jean-Yves Nief  For local organization & meeting web page Stephane Duray & administration team  For excellent logistical support 17