3rd Nov 2000HEPiX/HEPNT 20001 CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department

Slides:



Advertisements
Similar presentations
The Quantum Chromodynamics Grid James Perry, Andrew Jackson, Matthew Egbert, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Advertisements

Tableau Software Australia
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
SLC/Ver1.0/OS CONCEPTS/Oct'991INTRODUCTION What is an Operating System? Operating Structure -System Components -OS Services -System Calls & Programs -System.
Chapter 8 Operating System Support
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Operating Systems CS208. What is Operating System? It is a program. It is the first piece of software to run after the system boots. It coordinates the.
Computer Organization and Architecture
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
Distributed Systems: Client/Server Computing
11 Dec 2000F Harris Datagrid Testbed meeting at Milan 1 LHCb ‘use-case’ - distributed MC production
Test Review. What is the main advantage to using shadow copies?
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Overview of the Database Development Process
C Copyright © 2009, Oracle. All rights reserved. Appendix C: Service-Oriented Architectures.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
Operating Systems (CS 340 D) Dr. Abeer Mahmoud Princess Nora University Faculty of Computer & Information Systems Computer science Department.
ITEC224 Database Programming
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Recall: Three I/O Methods Synchronous: Wait for I/O operation to complete. Asynchronous: Post I/O request and switch to other work. DMA (Direct Memory.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Scalable Web Server on Heterogeneous Cluster CHEN Ge.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Nick Brook Current status Future Collaboration Plans Future UK plans.
SAM and D0 Grid Computing Igor Terekhov, FNAL/CD.
IT 456 Seminar 5 Dr Jeffrey A Robinson. Overview of Course Week 1 – Introduction Week 2 – Installation of SQL and management Tools Week 3 - Creating and.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 3: Operating-System Structures System Components Operating System Services.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
PLANNING ENGINEERING AND PROJECT MANAGEMENT By Lec. Junaid Arshad 1 Lecture#03 DEPARTMENT OF ENGINEERING MANAGEMENT.
Operating Systems David Goldschmidt, Ph.D. Computer Science The College of Saint Rose CIS 432.
Distributed DBMSs- Concept and Design Jing Luo CS 157B Dr. Lee Fall, 2003.
Metadata Mòrag Burgon-Lyon University of Glasgow.
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
UNIX Unit 1- Architecture of Unix - By Pratima.
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
BaBar and the GRID Tim Adye CLRC PP GRID Team Meeting 3rd May 2000.
A proposal: from CDR to CDH 1 Paolo Valente – INFN Roma [Acknowledgements to A. Di Girolamo] Liverpool, Aug. 2013NA62 collaboration meeting.
Adapting SAM for CDF Gabriele Garzoglio Fermilab/CD/CCF/MAP CHEP 2003.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
CLRC Grid Team Glenn Patrick LHCb GRID Plans Glenn Patrick LHCb has formed a GRID technical working group to co-ordinate practical Grid.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
1 Chapter 2: Operating-System Structures Services Interface provided to users & programmers –System calls (programmer access) –User level access to system.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
System Components Operating System Services System Calls.
Operating System Structure Lecture: - Operating System Concepts Lecturer: - Pooja Sharma Computer Science Department, Punjabi University, Patiala.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
LHCb computing model and the planned exploitation of the GRID Eric van Herwijnen, Frank Harris Monday, 17 July 2000.
File-System Management
Applied Operating System Concepts
File System Implementation
Multiple Processor Systems
Operating System Concepts
Chapter 2: Operating-System Structures
Chapter 2: Operating-System Structures
Short to middle term GRID deployment plan for LHCb
Operating System Concepts
Presentation transcript:

3rd Nov 2000HEPiX/HEPNT CDF-UK MINI-GRID Ian McArthur Oxford University, Physics Department

3rd Nov 2000HEPiX/HEPNT Background CDF collaborators in the UK applied for JIF grant for IT equipment in Awarded £1.67M in summer First half of grant will buy –Multiprocessor systems plus 1TB of disk for 4 Universities –2 multiprocessors plus 2.5 TB of disk for RAL –A 32 CPU farm for RAL –5 TB of disk and 8 high end workstations for FNAL Emphasis on high IO throughput ‘super- workstations’. A dedicated network link from London to FNAL

3rd Nov 2000HEPiX/HEPNT CDF-UK Equipment Bid

3rd Nov 2000HEPiX/HEPNT Hardware and Network Tender document is written and schedule is on target for equipment delivery in May Second phase starts June 2002 Developed a scheme for transparent access to CDF systems via the US link. –Each system CDF-UK requires to use the link has an alternative IP name and address to allow the data to be sent down the dedicated link. –A Network Address Translation scheme ensures that return traffic takes the same path (symmetric routing) –Demonstrated the scheme working with 2 Cisco routers on a local network. –Starting to talk to network providers to implement physical link. –Must try to make Kerberos work across this link

3rd Nov 2000HEPiX/HEPNT 20005

3rd Nov 2000HEPiX/HEPNT Software Project JIF proposal only covered hardware but in the meantime GRID has arrived ! Aim to provide a scheme to allow efficient use of the new equipment and other distributed resources. Concentrate on solving real-user issues. Develop an architecture for locating data, data transfer and job submission within a distributed environment Based on the GRID architecture initially on top of the Globus toolkit. Gives us experience in this rapidly developing field.

3rd Nov 2000HEPiX/HEPNT Some Requirements Want an efficient environment: so automate routine tasks as much as possible With few resources available must make best use of the existing packages and require few or no modifications to existing software. To make best use of the systems available: –data may need to be moved to where these is available CPU, –or a job may need to be submitted to a remote site to avoid moving the data. Produce a simple but useful system ASAP.

3rd Nov 2000HEPiX/HEPNT Design principles All sites are equal All sites hold meta-data describing only local data Use LDAP to publish meta-data kept in: –Oracle - at FNAL –msql - at most other places may go to MySQL –Can introduce caching but keep it simple at first Use local intelligence at each end of data transfer – allows us to take account of local idiosyncrasies e.g. use of near-line storage, disk space management Use existing Disk Inventory Manager

3rd Nov 2000HEPiX/HEPNT CDF Data Dataset: a primary dataset contains all the processed data from a specific physics channel. –Secondary datasets by event selection –Datasets will grow over time as more data is taken and data continues to be processed. Fileset: smallest collection of data which can be requested from the data handling system. At Fermilab, a fileset is mapped to a single partition on a tape and contains a few files. File: A member of a fileset. The smallest unit of data known to a filesystem, typically 1GB. Metadata: Stores relationships between files, filesets and datasets, run conditions, luminosity etc.

3rd Nov 2000HEPiX/HEPNT Data Location/Copy

3rd Nov 2000HEPiX/HEPNT Layers User Interface Dataset maintainer Data locator Data copier Globus toolkit... Job Submission

3rd Nov 2000HEPiX/HEPNT Functionality at a site A mechanism to allow jobs from participating sites to be run. Publication of the local metadata Publication of information about other system resources (CPU, Disk, Batch queues etc). Transmission of data via network. –This may involve staging of data from tape to disk before transmission. Receive data from the network or from tapes. Copy or construct metadata Some sites may have reduced functionality

3rd Nov 2000HEPiX/HEPNT Scope Plan to install at –4 UK universities (Glasgow, Liverpool, Oxford, UCL) –RAL –FNAL (although this would be reduced functionality, data and metadata exporter) –More non-UK sites could be included Intend to have basic utilities in place at time of equipment installation (May 2001)

3rd Nov 2000HEPiX/HEPNT Work so far Project plan under development – once finished additional resources will be requested. Globus installed at a number of sites. Remote execution of shell commands checked. Some bits demonstrated: –LDAP to Oracle via Python script Python convenient scripting language for the job May use a daemon to hold connection to ORACLE LDAP only implement search - and even this is quite tricky because your script should support filter, base and scope. LDAP schema will not reflect full SQL schema but just what is needed. –Java to LDAP (via JNDI) JNDI (Java Naming and Directory Interface) gives very elegant interface to LDAP

3rd Nov 2000HEPiX/HEPNT Longer Term Goals User Interface to be implemented as Java application to give platform independence. UI to automate or suggest strategies for moving data/submitting jobs –Need to include cost/elapsed time estimates for task completion –Need to look up dataset sizes, network health, time to copy from tape or disk, cpu load etc. Look for more generic solutions Evaluate any new GRID tools which might standardize any parts we’ve implemented ourselves. Consolidation with other GRID projects