Application Use Cases NIKHEF, Amsterdam, December 12, 13.

Slides:



Advertisements
Similar presentations
ESA Data Integration Application Open Grid Services for Earth Observation Luigi Fusco, Pedro Gonçalves.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
11 Decembre 2000V. Breton Milan WP6 DataGRID meeting Biological applications in testbed 0 Evaluate GRID added value for handling biological data –What.
The CrossGrid project Juha Alatalo Timo Koivusalo.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
EU 2nd Year Review – Jan – WP9 WP9 Earth Observation Applications Demonstration Pedro Goncalves :
EDG Application The European DataGrid Project Team
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
SLICE Simulation for LHCb and Integrated Control Environment Gennady Kuznetsov & Glenn Patrick (RAL) Cosener’s House Workshop 23 rd May 2002.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
DataGrid Applications Federico Carminati WP6 WorkShop December 11, 2000.
Nick Brook Current status Future Collaboration Plans Future UK plans.
A Grid Computing Use case Datagrid Jean-Marc Pierson.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Applications and Use Cases The European DataGrid Project Team
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
N ATIONAL E NERGY R ESEARCH S CIENTIFIC C OMPUTING C ENTER Charles Leggett The Athena Control Framework in Production, New Developments and Lessons Learned.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
4/20/02APS April Meeting1 Database Replication at Remote sites in PHENIX Indrani D. Ojha Vanderbilt University (for PHENIX Collaboration)
Event Data History David Adams BNL Atlas Software Week December 2001.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
28 March 2001F Harris LHCb Software Week1 Overview of GGF1 (Global Grid Forum) and Datagrid meeting, NIKHEF, Mar 5-9 F Harris(Oxford)
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
V. Breton, Budapest, September 2002 WP10 status V. Breton DataGrid conference.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
GDB Meeting - 10 June 2003 ATLAS Offline Software David R. Quarrie Lawrence Berkeley National Laboratory
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
EDG Applications The European DataGrid Project Team
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
MEDIGRID project, DataGrid FR meeting, April 18, 2002, Johan Montagnat, WP10 ACI GRID 2002 MEDIGRID: high performance medical image processing on a computational.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
EC Review – 01/03/2002 – WP9 – Earth Observation Applications – n° 1 WP9 Earth Observation Applications 1st Annual Review Report to the EU ESA, KNMI, IPSL,
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
2. WP9 – Earth Observation Applications ESA DataGrid Review Frascati, 10 June Welcome and introduction (15m) 2.WP9 – Earth Observation Applications.
INFSO-RI Enabling Grids for E-sciencE EGEE-2 NA4 Biomed Bioinformatics in CNRS Christophe Blanchet Institute of Biology and Chemistry.
- GMA Athena (24mar03 - CHEP La Jolla, CA) GMA Instrumentation of the Athena Framework using NetLogger Dan Gunter, Wim Lavrijsen,
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
VI/ CERN Dec 4 CMS Software Architecture vs Hybrid Store Vincenzo Innocente CMS Week CERN, Dec
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
EC Review – 01/03/2002 – F.Carminati – Accomplishments of the project from the end user point of view– n° 1 Accomplishments of the project from the end.
WP10 Goals and accomplishments from WP10 point of view J. Montagnat, CNRS, CREATIS V. Breton, CNRS/IN2P3 DataGrid Biomedical Work Package.
Applications and the Grid EDG CERN Ingo Augustin CERN DataGrid HEP Applications.
Earth Observation inputs to ATF Annalisa Terracina EU-DataGrid Project Work Package 9 – EO Applications April 2003 CERN.
DataGrid France 12 Feb – WP9 – n° 1 WP9 Earth Observation Applications.
WP 9.4 Use Case (ESA-ESRIN, IPSL, KNMI)
Moving the LHCb Monte Carlo production system to the GRID
EO Applications Parallel Session
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
US ATLAS Physics & Computing
Gridifying the LHCb Monte Carlo production system
ATLAS DC2 & Continuous production
Use Of GAUDI framework in Online Environment
Presentation transcript:

Application Use Cases NIKHEF, Amsterdam, December 12, 13

Use Cases part of a development process – requirements gathering – use cases – architectural design – fast prototyping – implementation text describing a real case

A D0 use case produce 1 million events using the Pythia event generator and the GEANT D0 detector simulation program. Add pile-up events during digitisation and before reconstruction. After reconstruction and analysis all raw and resulting data is conserved. Any production should be exactly reproducible.

Pythia GEANT-3 Simulation Reconstruction Analysis cards D0 MC Flow Chart Min. bias geometry INPUTOUTPUT

Requirements Gathering discussing use cases often delivers system requirements Design & Architecture new components extending existing components (standards!) CRC’s

Refinements # min.bias events depending on luminosity add 0< <10 min.bias events per event any event should be (exactly) reproducible min.bias events are also generated with Pythia min.bias events could also be measured data could be in a file of ## events could be stored one by one in a database

Options min.bias events generated on the fly min.bias events from a file on the grid min.bias events from file on the CE min.bias events from generated file etc. and always (exactly) reproducable !

HEPCAL Document requirements from all 4 LHC experiments requirements from EO and Bio ~50 use cases discussed within Architectural Task Force design is input for Global Grid Forum follow always standard protocols

ATLAS

An implementation of distributed analysis in ALICE using natural parallelism of processing Local Remote Selection Parameters Procedure Proc.C PROOF CPU TagD B RD B DB 1 DB 4 DB 5 DB 6 DB 3 DB 2 Bring the job to the data and not the data to the job

ATLAS/LHCb Software Framework (Based on Services) Converter Algorithm Event Data Service Persistency Service Data Files Algorithm Transient Event Store Detec. Data Service Persistency Service Data Files Transient Detector Store Message Service JobOptions Service Particle Prop. Service Other Services Histogram Service Persistency Service Data Files Transient Histogram Store Application Manager Converter The Gaudi/Athena Framework – Services will interface to Grid (e.g. Persistency)

A CMS Data Grid Job The vision for 2003

Common Applications Work Several discussions between application WPMs and technical coordination to consider the common needs of all applications HEPEOBio Common applicative layer EDG software Globus

reconstruction simulation analysis interactive physics analysis batch physics analysis batch physics analysis detector event summary data raw data event reprocessing event reprocessing event simulation event simulation analysis objects (extracted by physics topic) Data Handling and Computation for Physics Analysis event filter (selection & reconstruction) event filter (selection & reconstruction) processed data CER N

LCG/Pool on the Grid File Catalog Collections Replica Location Service Grid Dataset Registry Grid Resources Experiment Framework User Application LCG POOLGrid Middleware RootI/O Replica Manager

Applications in DataGrid HEP Bio Informatics and Health Earth Observation

Challenges for a biomedical grid The biomedical community has NO strong center of gravity in Europe –No equivalent of CERN (High-Energy Physics) or ESA (Earth Observation) –Many high-level laboratories of comparable size and influence without a practical activity backbone (EMB-net, national centers,…) leading to: Little awareness of common needs Few common standards Small common long-term investment The biomedical community is very large (tens of thousands of potential users) The biomedical community is often distant from computer science issues

Biomedical requirements Large user community(thousands of users) –anonymous/group login Data management –data updates and data versioning –Large volume management (a hospital can accumulate TBs of images in a year) Security –disk / network encryption Limited response time –fast queues High priority jobs –privileged users Interactivity –communication between user interface and computation Parallelization –MPI site-wide / grid-wide –Thousands of images –Operated on by 10’s of algorithms Pipeline processing –pipeline description language / scheduling

Biomedical projects in DataGrid Distributed Algorithms. New distributed "grid-aware" algorithms (bio-info algorithms, data mining, …) Grid Service Portals. Service providers taking advantage of the DataGrid computational power and storage capacity. Cooperative Framework. Use the DataGrid as a cooperative framework for sharing resources, algorithms, and organize experiments in a cooperative manner. Cooperative Framework Grid Service Portals Distributed Algorithms EDG Middleware WP10 applications

The grid impact on data handling DataGrid will allow mirroring of databases –An alternative to the current costly replication mechanism –Allowing web portals on the grid to access updated databases Biomedical Replica Catalog Trembl(EBI) Swissprot (Geneva)

Web portals for biologists Biologist enters sequences through web interface Pipelined execution of bio-informatics algorithms –Genomics comparative analysis (thousands of files of ~Gbyte) Genome comparison takes days of CPU (~n**2) –Phylogenetics –2D, 3D molecular structure of proteins… The algorithms are currently executed on a local cluster –Big labs have big clusters … –But growing pressure on resources – Grid will help More and more biologists compare larger and larger sequences (whole genomes)… to more and more genomes… with fancier and fancier algorithms !!

The Visual DataGrid Blast, a first genomics application on DataGrid A graphical interface to enter query sequences and select the reference database A script to execute the BLAST algorithm on the grid A graphical interface to analyze result Accessible from the web portal genius.ct.infn.it

Summary of added value provided by Grid for BioMed applications Data mining on genomics databases (exponential growth). Indexing of medical databases (Tb/hospital/year). Collaborative framework for large scale experiments (e.g. epidemiological studies). Parallel processing for –Databases analysis –Complex 3D modelling

Earth Observation (WP9)  Global Ozone (GOME) Satellite Data Processing and Validation by KNMI, IPSL and ESA  The DataGrid testbed provides a collaborative processing environment for 3 geographically distributed EO sites (Holland, France, Italy) 28