Sergiu January 2007 TG Users’ Data Transfer Needs SDSC NCAR TACC UC/ANL NCSA ORNL PU IU PSC.

Slides:



Advertisements
Similar presentations
Cross-site data transfer on TeraGrid using GridFTP TeraGrid06 Institute User Introduction to TeraGrid June 12 th by Krishna Muriki
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
TeraGrid Quarterly Meeting Dec 6-7, 2007 DVS GIG Project Year 4&5 Project List Kelly Gaither, DVS Area Director.
XSEDE 13 July 24, Galaxy Team: PSC Team:
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 5: Managing File Access.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 5: Managing File Access.
Hands-On Microsoft Windows Server 2003 Networking Chapter 7 Windows Internet Naming Service.
Common System Components
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
SM3121 Software Technology Mark Green School of Creative Media.
Simo Niskala Teemu Pasanen
Cambodia-India Entrepreneurship Development Centre - : :.... :-:-
Network, Operations and Security Area Tony Rimovsky NOS Area Director
GridFTP Guy Warner, NeSC Training.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
SSO current status 10/6/10 Area Director’s call. Easy as 1-2-3! Fully diagrammed login and certificate set-up process, pre- Single Sign-on You can see.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 5: Managing File Access.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Problem Determination Your mind is your most important tool!
1 G A A new Document Control System “A new system to manage LIGO documents” Stuart Anderson Melody Araya David Shoemaker 29 September, 2008
The iPlant Collaborative Community Cyberinfrastructure for Life Science Tools and Services Workshop Objectives.
Algoval: Evaluation Server Past, Present and Future Simon Lucas Computer Science Dept Essex University 25 January, 2002.
Accelerating Scientific Exploration Using Workflow Automation Systems Terence Critchlow (LLNL) Ilkay Altintas (SDSC) Scott Klasky(ORNL) Mladen Vouk (NCSU)
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 3: Operating-System Structures System Components Operating System Services.
INFSO-RI Enabling Grids for E-sciencE DAGs with data placement nodes: the “shish-kebab” jobs Francesco Prelz Enzo Martelli INFN.
Group 1 : Grid Computing Laboratory of Information Technology Supervisors: Alexander Ujhinsky Nikolay Kutovskiy.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Page 1 Printing & Terminal Services Lecture 8 Hassan Shuja 11/16/2004.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Sergiu April 2006June 2006 Overview of TeraGrid Resources and Services Sergiu Sanielevici, TeraGrid Area Director for User.
Distributed File Systems 11.2Process SaiRaj Bharath Yalamanchili.
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
Cloud Computing Computer Science Innovations, LLC.
Data, Visualization and Scheduling (DVS) TeraGrid Annual Meeting, April 2008 Kelly Gaither, GIG Area Director DVS.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
File Transfer And Access (FTP, TFTP, NFS). Remote File Access, Transfer and Storage Networks For different goals variety of approaches to remote file.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Holding slide prior to starting show. Lessons Learned from the GECEM Portal David Walker Cardiff University
Climate-SDM (1) Climate analysis use case –Described by: Marcia Branstetter Use case description –Data obtained from ESG –Using a sequence steps in analysis,
GridShell/Condor: A virtual login Shell for the NSF TeraGrid (How do you run a million jobs on the NSF TeraGrid?) The University of Texas at Austin.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
This was written with the assumption that workbooks would be added. Even if these are not introduced until later, the same basic ideas apply Hopefully.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
TeraGrid Software Integration: Area Overview (detailed in 2007 Annual Report Section 3) Lee Liming, JP Navarro TeraGrid Annual Project Review April, 2008.
Tamas Kiss University Of Westminster
Java Beans Sagun Dhakhwa.
Thoughts on Computing Upgrade Activities
US CMS Testbed.
Cyberinfrastructure and PolarGrid
Chapter 2: Operating-System Structures
Overview of Workflows: Why Use Them?
Java Remote Method Invocation
Chapter 2: Operating-System Structures
Presentation transcript:

Sergiu January 2007 TG Users’ Data Transfer Needs SDSC NCAR TACC UC/ANL NCSA ORNL PU IU PSC

Sergiu January 2007 Agents provocateurs I will kick off a discussion based on two use cases. Other user support folks in the audience will then chime in. This should give you a good sampling of user requirements, since all users basically want the same things and don’t really care how it’s done.

Sergiu January 2007 GASOLINE N-body cosmology software system by T. Quinn et al., U. Washington Phase 1: Generate Initial Conditions (ICs) –On Cobalt at NCSA: Job 1: generate 3 “force files” o(1-100 GB) Interactive GridFTP transfer of force files to PSC archive Job 2: from force files to IC file o( 100 MB – 100 GB) GridFTP transfer of IC file to PSC BigBenGridFTP transfer of IC file to PSC BigBen Phase 2: Simulation –On BigBen at PSC: Run for several weeks per IC. “Small cases”: output data files o(100 MB), total for run o(100 GB) “Large cases”: output data files o(100 GB), total for run o(2 TB) Archive at PSC (in future maybe also elsewhere on TG)

Sergiu January 2007 GASOLINE cont’d Phase 3: Data analysis –“Small” files (~100 MB each) directly to UW –“Large files” (~100 GB each) to Cobalt at NCSA Use GridFTP, need each transfer to complete in ~hrsUse GridFTP, need each transfer to complete in ~hrs Phase 4: Visualization using SALSA –Done at UW for “small” files –For “large” files TCS or NCSA DTF (XT3 port underway) Need distributed memory machine that can handle dynamic linking and whose compute nodes have sockets connectivity to a machine that in turn can communicate with the client application at UW.

Sergiu January 2007 GASOLINE Wish List Reliability –“We have a very low success rate in being able to initiate a GridFTP file transfer. Usually there is some authentication issue that appears either during "grid- proxy-init" or during the invocation of the GridFTP client itself.” Batch-initiated file transfers (BIFT) – Enable spawning a GridFTP file transfer at the end of a compute job. E.g. when an IC file is generated on Cobalt, spawn a GridFTP transfer to the PSC archiver. When a simulation job completes on BigBen, spawn a GridFTP transfer of the new data files to the reduction or visualization system.

Sergiu January 2007 Wish list con’t More on BIFT: –“Whether this is done via an explicit command-line transfer or a WAN-FS is not important – we just need near 100% reliability, and transfer to occur even several days after the original PBS job was submitted. The limited lifetime of proxies usually makes this difficult. –Please note that we do not want to do the file transfer as part of the actual compute job. We want to be able to dump the data during the simulation to the fastest local file system that is available. When the compute job completes, it should spawn a separate job (for which we are not billed 2048 SUs per hour) that does the file transfer.”

Sergiu January 2007 One more thing… “One thing I would highly recommend for command line file transfers is that the command line be somewhat shorter than globus-url-copy requires. Every researcher that I talk to about globus-url-copy does not use it because too much typing is involved. tgcp is supposed to address this problem, but it I have found it to be extremely unreliable at this point. Usually its "translation" from tgcp format to globus-url-copy format is incorrect and its call to globus-url-copy does not work. The best idea, I think, would be if someone could make a GridFTP client that was invoked using exactly the same conventions as "scp." Researchers would not have to learn anything new to be able to use it.” to use it.” –Jeff Gardner, PSC, Gasoline co-author & team member.

Sergiu January 2007 How to get MyCluster users to transfer files via TG? “Most users of MyCluster actually only use it to submit to a single machine. No matter how much I show them MyCluster’s amazing ability to easily run on multiple systems (which you can accomplish by adding just a single line to a configuration file), they have no interest in using this ability since they do not want to deal with file transfers between TeraGrid sites. Although they seem to understand that the automatic file transfers are actually quite simple to set up in MyCluster, they still resist the idea. Thus, the obstacle seems simply to be the concept of having their files automatically transferred to and from a location. They are not used to having this capability, so they avoid it. On the other hand, several have asked about shared file systems, and a few have indicated that if there were indeed some shared file system between multiple TeraGrid sites, that would make the concept of distributing their runs across sites more palatable. Most users have experience with NFS. Since they can directly map this onto something they already use, they seem much more comfortable with the idea.” On the other hand, several have asked about shared file systems, and a few have indicated that if there were indeed some shared file system between multiple TeraGrid sites, that would make the concept of distributing their runs across sites more palatable. Most users have experience with NFS. Since they can directly map this onto something they already use, they seem much more comfortable with the idea.” – Jeff Gardner, Ed Walker’s partner in MyCluster testing & deployment.

Sergiu January 2007 My Summary – Start of Debate In the TG system, there are two kinds of users: ftp’ers and NFSniks. Both need reliability first and foremost.Both need reliability first and foremost. Both care about performance, in the sense that files must be where needed when needed – the “when” can vary from minutes to ~ 1dayBoth care about performance, in the sense that files must be where needed when needed – the “when” can vary from minutes to ~ 1day But ftp’ers tend to care more (that’s why they’ve learned ftp/scp in the first place) User interface and scripting support are essential for tool adoption.User interface and scripting support are essential for tool adoption. Persistent WAN-FS hosting of community code as well asPersistent WAN-FS hosting of community code as well as working datasets would enhance many user groups’ productivity.

Sergiu January 2007 Looking ahead TB-months disk storage of working datasets is increasingly important as TG fosters complex WAN workflows.TB-months disk storage of working datasets is increasingly important as TG fosters complex WAN workflows. TG WIDE and OPEN will also place increasing importance on read-only access to Archives and Repositories.TG WIDE and OPEN will also place increasing importance on read-only access to Archives and Repositories. Live instruments and Repositories to be integrated with computational workflows => complex data workflows, cf. Tier 0 – 2 for LHC experiments (OSG)Live instruments and Repositories to be integrated with computational workflows => complex data workflows, cf. Tier 0 – 2 for LHC experiments (OSG) Track 1 and Track 2 systems will be, in essence, large instruments…