Marc Mengel Solution for the last mile - IFDH. Marc Mengel IFDH Thin Layer for Data Movement Automatic protocol discovery or specify with environment.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
TOPIC FOR THE DISCUSSION SECTION AUTHOR ONE AUTHOR TWO AUTHOR THREE FIFE Workshop Template Service Name.
EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
GridFTP: File Transfer Protocol in Grid Computing Networks
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Minerva Infrastructure Meeting – October 04, 2011.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Operating Systems.  Operating System Support Operating System Support  OS As User/Computer Interface OS As User/Computer Interface  OS As Resource.
Thank You ©2012, Cognizant. Rapido has been created by the Research and Development team from QE&A Technology CoE Rapido is continuously enhanced and.
Security Encryption Authentication Usability Command Line History Output and Input clear Scheduling Integration with existing scheduling tools Separate.
DPM CCRC - 1 Research and developments DPM status and plans Jean-Philippe Baud.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Through the development of advanced middleware, Grid computing has evolved to a mature technology in which scientists and researchers can leverage to gain.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
The Network Performance Advisor J. W. Ferguson NLANR/DAST & NCSA.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
OSG Storage Architectures Tuesday Afternoon Brian Bockelman, OSG Staff University of Nebraska-Lincoln.
Ganga A quick tutorial Asterios Katsifodimos Trainer, University of Cyprus Nicosia, Feb 16, 2009.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Archiving Scientific Data Outside of the Traditional HEP Domain, Using the Archive Facilities at Fermilab M. Diesburg, M. Gheith, R. Illingworth, M. Mengel,
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Storage and Data Movement at FNAL D. Petravick CHEP 2003.
Virtual Data Management for CMS Simulation Production A GriPhyN Prototype.
Status Report on the Validation Framework S. Banerjee, D. Elvira, H. Wenzel, J. Yarba Fermilab 15th Geant4 Collaboration Workshop 10/06/
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI How to integrate portals with the EGI monitoring system Dusan Vudragovic.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
TANYA LEVSHINA Monitoring, Diagnostics and Accounting.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Archiving Scientific Data Outside of the Traditional HEP Domain, Using the Archive Facilities at Fermilab M. Diesburg, M. Gheith, R. Illingworth, M. Mengel,
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Parag Mhashilkar (Fermi National Accelerator Laboratory)
SAM/DAG/Job Submission Integration Dennis Box, FNAL REX/CD 8/14/12.
Simulation Production System
Data services on the NGS
Gfal/lcg-util -> Gfal2/gfal2-util
Job workflow Pre production operations:
GFAL 2.0 Devresse Adrien CERN lcgutil team
Hironori Ito Brookhaven National Laboratory
Viet Tran Institute of Informatics Slovakia
Interoperability & Standards
Large Scale Test of a storage solution based on an Industry Standard
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
CUPS Print Services.
Andy Puckett – Sales Engineer
Mats Rynge USC Information Sciences Institute
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Marc Mengel Solution for the last mile - IFDH

Marc Mengel IFDH Thin Layer for Data Movement Automatic protocol discovery or specify with environment var Protocols supported: cp (throttled), srm, gridftp, xrootd (near future)

Marc Mengel Users need not worry about protocol Users’ job scripts use ifdh for transfers - simple while read sourcefile do ifdh cp $sourcefile localsource framework_exe -c config localsource localout ifdh cp localout $outarea done < playlist With a data handling system while uri=`ifdh getNextFile $projectUrl` && [-n “$uri”] do localsource = `ifdh fetchInput $uri` framework_exe -c config $localsource localout ifdh cp localout $outarea done With an integrated framework framework_exe -c config --samProjectUrl $projectUrl -o $outarea ifdh copyBackOutput $myOutArea $outarea

Marc Mengel Protect the central storage ifdh cp to central storage is throttled by a system called “CPN” Allows n simultaneous transfers per experiment to Bluearc Transfers exceeding are queued The challenge is to ensure that all users utilize CPN. ifdh cp makes that easy Interactive Nodes Central Storage (NFS) BatchNodes

Marc Mengel Harder part – transferring output back home Direct Worker Node Node dropbox dropbox Node Storage Element Worker Node Node Best man srm dropbox srm Fermilab Offsite Fermilab Offsite Direct via SRM 3rd party SRM

Marc Mengel Benefits of the thin abstraction Users are shielded from details of protocol choice We can change the protocols and decision algorithms without breaking the user’s scripts Other features: o Tools to define and view SAM data-sets o Tools to locate files on tape or in cache o Logging over UDP to monitoring services o Supports many languages: C++, Python, Bash

Marc Mengel Examples #include “ifdh.h” ifdh i(); location = i.locateFile(base_uri, filename); import ifdh i = ifdh.ifdh() location = i.locateFile(base_uri, filename) location = `ifdh localFile $base_uri $filename` C++ Python Bash shell Easy to add more since we use SWIG

Marc Mengel New Features: ● Shared/Auxiliary file interface: fetchSharedFiles findMatchingFiles ● Utility Functions ● ls/lss/ll ● chmod ● rename ● more