How to Install and Use the DQ2 User Tools US ATLAS Tier2 workshop at IU June 20, 2007 - Bloomington, IN Marco Mambelli University of Chicago.

Slides:



Advertisements
Similar presentations
LNL M.Biasotto, Roma, 22 novembre Some notes on WP6 INFN Testbed.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Grid data transfer tools Tier 3(g,w) meeting at ANL ASC – May 19, 2009 Marco Mambelli – University of Chicago
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Ninth EELA Tutorial for Users and Managers E-infrastructure shared between Europe and Latin America LFC Server Installation and Configuration.
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
Data management at T3s Hironori Ito Brookhaven National Laboratory.
OSG Services at Tier2 Centers Rob Gardner University of Chicago WLCG Tier2 Workshop CERN June 12-14, 2006.
CERN Manual Installation of a UI – Oxford July - 1 LCG2 Administrator’s Course Oxford University, 19 th – 21 st July Developed.
1 DDM Troubleshooting Console (quick evaluation of splunk for such) Rob Gardner University of Chicago ATLAS DDM Workshop January 26, 2007.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
DDM-Panda Issues Kaushik De University of Texas At Arlington DDM Workshop, BNL September 29, 2006.
Tier 3 Storage Survey and Status of Tools for Tier 3 access, futures Doug Benjamin Duke University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
BNL DDM Status Report Hironori Ito Brookhaven National Laboratory.
Open Science Grid OSG CE Quick Install Guide Siddhartha E.S University of Florida.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
June 22, 2007USATLAS T2-T3 DQ2 0.3 SiteServices Patrick McGuigan
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Nurcan Ozturk University of Texas at Arlington US ATLAS Transparent Distributed Facility Workshop University of North Carolina - March 4, 2008 A Distributed.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
The ATLAS Cloud Model Simone Campana. LCG sites and ATLAS sites LCG counts almost 200 sites. –Almost all of them support the ATLAS VO. –The ATLAS production.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
EGEE is a project funded by the European Union under contract IST VO box: Experiment requirements and LCG prototype Operations.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Service Availability Monitor tests for ATLAS Current Status Tests in development To Do Alessandro Di Girolamo CERN IT/PSS-ED.
Distributed Data Management Miguel Branco 1 DQ2 discussion on future features BNL workshop October 4, 2007.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
Open Science Grid Build a Grid Session Siddhartha E.S University of Florida.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
Tier 3 Support and the OSG US ATLAS Tier2/Tier3 Workshop at UChicago August 20, 2009 Marco Mambelli –
Distributed Data Management Miguel Branco DDM development Architecture and developers workplan on behalf of the DDM group.
The GridPP DIRAC project DIRAC for non-LHC communities.
12th EELA Tutorial for Users and Managers E-infrastructure shared between Europe and Latin America LFC Server Installation and Configuration.
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Discussion on data transfer options for Tier 3 Tier 3(g,w) meeting at ANL ASC – May 19, 2009 Marco Mambelli – University of Chicago
Data Management at Tier-1 and Tier-2 Centers Hironori Ito Brookhaven National Laboratory US ATLAS Tier-2/Tier-3/OSG meeting March 2010.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
VO Box discussion ATLAS NIKHEF January, 2006 Miguel Branco -
CMS User Support and Beijing Site Xiaomei Zhang CMS IHEP Group Meeting March
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
1 DIRAC Project Status A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 10 March, DIRAC Developer meeting.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America LFC Server Installation and Configuration.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
gLite Data management system overview
Extended OSG client for WLCG
Hands-On Session: Data Management
Data services in gLite “s” gLite and LCG.
Architecture of the gLite Data Management System
Site availability Dec. 19 th 2006
Presentation transcript:

How to Install and Use the DQ2 User Tools US ATLAS Tier2 workshop at IU June 20, Bloomington, IN Marco Mambelli University of Chicago

Topics No grid-setup No certificate/proxy DQ2 Move data between local SE, Tier1/2, Tier3 and local disks Install/configure user tools Use user tools

Some technology 3 Grids in ATLAS (OSG, LCG, NG):  Different file catalogs: Local Replica Catalog (LRC) - is used in OSG Local File Catalog (LFC) - is used in LCG  Different information system  Different environment Software  Mostly python Platform independent This would be OK if only OSG/NG  LCG UI Os dependent Necessary to interact with LCG

DDM Overview DQ2 has four main components:  Central Dataset Catalogues (CERN) responsible for bookkeeping information  Site services (Tier1/2) responsible for fulfilling data movement requests (subscriptions)  Client tools to interact with central dataset catalogues and site services.  The LCG/ARDA Dashboard for monitoring data movement requests ement ement icesP1 icesP1

Topology example Tier0/1 Tier2/3 FTS Local/remote connection Local/remote DQ2 (different transfer protocols) CERN SE CE SE CE SE CE Tier1 SE

Site services Subscriptions work only between sites with DQ2 Site Services DQ2 Will be available at Tier1 and Tier2s  Patrick (Friday) Tier3 may install it Tier3 can use the DQ2 service of a ‘neighboring’ Tier2  Local files not in the catalog (2 step transfers)  Integrated catalog and SE (LRC required)  Integrated SE (donate their SE to the Tier2) siteservices/html/admin/ siteservices/html/admin/

DQ2 Clients Necessary to access DQ2 data Available for local installation or through AFS DQ2 (Catalog) Client  Allows to manage subscriptions (transfer files between DQ2 sites)  DQ2 User Tools  Allows to transfer files and datasets to local/remote directories  Data Moving Utilities (DMU), used by Panda and DDS

Catalog Clients Use from AFS (For latest production release -0.3.x):  Setup: /afs/cern.ch/atlas/offline/external/GRID/ddm/pro03/dq2.sh /afs/usatlas.bnl.gov/Grid/Don-Quijote/DQ2_0_3_client/setup.sh Installation (0.2 ?):  Probably just install a copy of the directory tree and fix the setup file dq2-register-subscription -a users.MarcoMambelli.testdataset1 UC

User Tools Access through AFS (source configuration file):  Setup LCG UI (or a grid environment) NameError: global name 'lfc' is not defined The lfc module is required to access LCG datasets. It is available in LCG UI or RPM. Then, lfc.py needs to be in one of the directories in PYTHONPATH. LFC-interfaces (SL3, SL4 as well):  Different setup: setup.[c,z]sh.[CERN,BNL,any]: e.g. setup.sh.any /afs/cern.ch/atlas/offline/external/GRID/ddm/endusers/ /afs/usatlas.bnl.gov/Grid/Don-Quijote/dq2_user_client/ Download from the CVS repository (Web accesible):  atlas.cgi/offline/DataManagement/DQ2_0_2/endusers/?only_with _tag=MAIN atlas.cgi/offline/DataManagement/DQ2_0_2/endusers/?only_with _tag=MAIN Doc:

User Tools Setup DQ2_URL_SERVER, DQ2 server $ export DQ2_URL_SERVER= DQ2_URL_SERVER_SSL, DQ2 server for secure connection $ export DQ2_URL_SERVER_SSL= DQ2_LOCAL_ID, local DQ2 site ID (from TiersOfATLAS) Depends on where you (e.g. CERN/lxplus, BNL users) $ export DQ2_LOCAL_ID=CERN $ export DQ2_LOCAL_ID=BNL  For other site users, ask your DDM admin for the ID. If your site doesn't deploy a DQ2 site service, leave it blank $ export DQ2_LOCAL_ID= DQ2_LOCAL_PROTOCOL, Protocol to access the local storage (rfio,castor,dcap,unix,dpm). (e.g. For CERN/lxplus, BNL users) $ export DQ2_LOCAL_PROTOCOL=castor $ export DQ2_LOCAL_PROTOCOL=dcap  If you use normal disk storage: $ export DQ2_LOCAL_PROTOCOL=unix

User Tool Setup (cont) DQ2_STORAGE_ROOT, root directory of local storage $ export DQ2_STORAGE_ROOT=/pfns DQ2_SRM_HOST, local SRM server (if any) $ export DQ2_SRM_HOST=srm://castorgrid.cern.ch:8443 DQ2_GSIFTP_HOST, local GridFTP server (if any) $ export DQ2_GSIFTP_HOST=gsiftp://castorgrid.cern.ch:2811 DQ2_USE_SRM, use SRM for all data transfer (default: False) $ export DQ2_USE_SRM=True LCG_CATALOG_TYPE, LCG catalog type $ export LCG_CATALOG_TYPE=lfc DQ2_LFC_HOME, LCG_HOME of local replica catalog $ export DQ2_LFC_HOME=/grid/atlas DQ2_COPY_COMMAND, which command is called in dq2_get. Specify this when srmcp doesn't work in your environment $ export DQ2_COPY_COMMAND='lcg-cp -v --vo atlas'

Available tools dq2_get  If a dataset is already present on a local storage element (SE) data is copied to the local directory or to another directory in the SE. dq2_ls  List information about datasets matching a given pattern. The pattern may contain wildcards which represent any strings. The wildcard symbol is the asterisk *. dq2_poolFCjobO  dq2_poolFCjobO resolves GUIDs for constituent files of a DQ2 dataset, and creates PoolFileCatalog.xml and Athena job-option.

Available tools (cont) dq2_put  Registers files to LRC, creates a dataset which is composed of the files, and then registers the dataset to DQ2. If a PoolFileCatalog is given, a list of files is extracted from the PoolFileCatalog. Otherwise, GUIDs are generated for files under a directory using uuidgen. dq2_register  It will upload local files to the Grid storage at the specified site and register them in the local Grid catalog and in DQ2 central catalogs.

Monitoring Panda dashboard:  mon/query?dash=ddm mon/query?dash=ddm ARDA dashboard:  test.cern.ch/dashboard/request.py/site test.cern.ch/dashboard/request.py/site

Reference User Tools Twiki  CVS  atlas.cgi/offline/DataManagement/DQ2_0_2/endusers/?only_with _tag=HEAD atlas.cgi/offline/DataManagement/DQ2_0_2/endusers/?only_with _tag=HEAD  (but check also - still empty) bin/viewcvs- atlas.cgi/offline/DataManagement/DQ2/dq2.enduser/?only_with_t ag=HEADhttp://atlas-sw.cern.ch/cgi- bin/viewcvs- atlas.cgi/offline/DataManagement/DQ2/dq2.enduser/?only_with_t ag=HEAD Generic DQ2  ent ent Help (bug submission, troubleshooting)  