Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.

Slides:



Advertisements
Similar presentations
Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Advertisements

LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Service Challenge Meeting “Service Challenge 2 update” James Casey, IT-GD, CERN IN2P3, Lyon, 15 March 2005.
Technical Architectures
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Summary of issues and questions raised. FTS workshop for experiment integrators Summary of use  Generally positive response on current state!  Now the.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Author: Andrew C. Smith Abstract: LHCb's participation in LCG's Service Challenge 3 involves testing the bulk data transfer infrastructure developed to.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MSG - A messaging system for efficient and.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Data management demonstrators Ian Bird; WLCG MB 18 th January 2011.
LCG Service Challenges SC2 Goals Jamie Shiers, CERN-IT-GD 24 February 2005.
LCG Storage Workshop “Service Challenge 2 Review” James Casey, IT-GD, CERN CERN, 5th April 2005.
Service Challenge Meeting “Review of Service Challenge 1” James Casey, IT-GD, CERN RAL, 26 January 2005.
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Baseline Services Group Status of File Transfer Service discussions Storage Management Workshop 6 th April 2005 Ian Bird IT/GD.
EGEE Data Management Services
Simulation Production System
Real Time Fake Analysis at PIC
“A Data Movement Service for the LHC”
Dynamic Deployment of VO Specific Condor Scheduler using GT4
IT-DB Physics Services Planning for LHC start-up
LCG 3D Distributed Deployment of Databases
Service Challenge 3 CERN
Grid Deployment Area Status Report
BNL FTS services Hironori Ito.
Sergio Fantinel, INFN LNL/PD
Castor services at the Tier-0
Data Management cluster summary
lundi 25 février 2019 FTS configuration
The LHCb Computing Data Challenge DC06
Presentation transcript:

Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005

Introduction Clarify what is expected to be in place to support the service challenges Where are the responsibilities Three components: Mass storage systems and their grid interfaces Data transfer infrastructure at a site File transfer service

Mass storage systems Interface to storage systems is SRM Tier 1: Local integrations of MSS with SRM (e.g. CASTOR, dCache, etc) Large Tier 2: Either as Tier 1, or large disk pools managed by dCache (or other SRM-enabled cache manager) Entirely the responsibility of each site to arrange for this integration

Infrastructure at a site What is needed to support the data transfer rates anticipated Scaleable transfer service infrastructure – disk servers, gridftp servers etc Again – clearly the responsibility of each Tier 1 to ensure this is in place in a timely fashion Examples: CERN infrastructure Ideas for SAN-based infrastructure – being prototyped with Taipei

Tier-0 Configuration 10 dual CPU Itanium machines Each has 1Gb connectivity that gets aggregated into a 10Gb switch All nodes run CERN SLC3 ~200GB local fast disk – 50MB/s single stream per disk Configured to run the following services: 5 gridftp servers, non-load balanced A 4-node load-balanced SRM/Gridftp system 1 control node, for running the transfer management software Direct connections to external network 10 Gb connection to GEANT 10 Gb link to Chicago (via Starlight) 10 Gb test link to SARA/NIKHEF (via Surfnet)

Prototypical SAN at a Tier 1

Site infrastructure – 2 Must foresee 2 sets of data transfer services Service challenge  next step up in performance Ongoing service  as each step is demonstrated in the service challenge it should be moved into production The production service continually improves in performance Typically at present Expect ~10 transfer nodes to achieve 500 MB/s Multiple requests, multiple files, multiple streams, …

File transfer service software Architecture designed with gLite Data Management team Proposal produced July 2004 Uses a loosely coupled agent-based model to allow for interaction of different components Current Solution Transfers carried out by simple perl-based multi-process daemon Simply wraps globus-url-copy UI/ Web Services submission interface from gLite User interface already what gLite users will see Integrated gLite FTS by Spring 2005 As the software becomes available, we will try it We need it for the March 2005 Service Challenge Good testbed for gLite – currently separate from rest of LCG-2 infrastructure  This will be a joint development between GDA and gLite teams

Transfer software (radiant ) Movement Scheduler: takes requests from queue and schedules according to a policy Uses: Movement Client to run and monitor the transfer; reschedules after failures Statistics Gatherer to store stats and performance info Request Store: Stores all requests submitted Tracks state of request Data Transport Server: Responsible for actual data transfer Runs on node where data resides E.g. SRM/Gridftp servers Agents: Autonomous – communicate through request store Can act on transfers in a given state Allows a VO to add actions (e.g. catalog registration)

Project Management Personnel Jamie Shiers joins Grid Deployment team at CERN with responsibility for Service Challenges Communication Website created: Mailing list for CERN service team created –Single point of contact for any issues related to CERN side of service challenge –List used for dissemination of general information and discussion with Tier-1s about current results and technical issues –Does this need to split into two – technical and “info” ?

Current Site Progress (Dec 04) RALFermilabBrookhavenKarlsruheIN2P3CNAFPIC 1. Network ConfigurationContact  2. Storage Configation  3. Site Tuning  4. “Service Challenge” Dec 04Jan 05 TaipeiNordugridTRIUMFNLDESY 1. Network Configuration Contact  2. Storage Configation  3. Site Tuning  4. “Service Challenge” Dec 04 KEY  = Complete  = In Progress

Summary Tier 1 sites must plan and set up this infrastructure rapidly now Network and Storage setup each take ~ 2 weeks to complete Tuning can take much longer Difficult to work out which layer is causing the problem Systems don’t scale linear with number of nodes Service Challenges require scheduling of many resources Personnel Network allocation Storage allocation

Ongoing Data Transfers * Graphs produced 01/12/04