11/01/20081 Data simulator status CCRC’08 Preparatory Meeting Radu Stoica, CERN* 11 th January 2007 * On leave from IFIN-HH.

Slides:



Advertisements
Similar presentations
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Online Magazine Bryan Ng. Goal of the Project Product Dynamic Content Easy Administration Development Layered Architecture Object Oriented Adaptive to.
Hands-On Microsoft Windows Server 2003 Administration Chapter 6 Managing Printers, Publishing, Auditing, and Desk Resources.
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Control and monitoring of on-line trigger algorithms using a SCADA system Eric van Herwijnen Wednesday 15 th February 2006.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
CLEO’s User Centric Data Access System Christopher D. Jones Cornell University.
Module 10 Configuring and Managing Storage Technologies.
Copyright © 2003 by Prentice Hall Module 4 Database Management Systems 1.What is a database? Data hierarchy and data organization Field, record, file,
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Hosted by Designing a Backup Architecture That Actually Works W. Curtis Preston President/CEO The Storage Group.
Introduction: Distributed POOL File Access Elizabeth Gallas - Oxford – September 16, 2009 Offline Database Meeting.
Marianne BargiottiBK Workshop – CERN - 6/12/ Bookkeeping Meta Data catalogue: present status Marianne Bargiotti CERN.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
What is expected from ALICE during CCRC’08 in February.
The european ITM Task Force data structure F. Imbeaux.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
MINER A Software The Goals Software being developed have to be portable maintainable over the expected lifetime of the experiment extensible accessible.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
 Database Administration Installing Oracle 11g & Creating Database.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
Grid Deployment Enabling Grids for E-sciencE BDII 2171 LDAP 2172 LDAP 2173 LDAP 2170 Port Fwd Update DB & Modify DB 2170 Port.
Page 1 Printing & Terminal Services Lecture 8 Hassan Shuja 11/16/2004.
Online System Status LHCb Week Beat Jost / Cern 9 June 2015.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Data management in LCG and EGEE David Smith.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Online/offline integration - Aart Heijboer - sector/soft - Cern june 2002 Integration Online and Offline software Run Control Ligier (dispatcher) Data.
20 Copyright © 2008, Oracle. All rights reserved. Cache Management.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
1 CTP offline software status (Offline week,8/4/08) R.Lietava for CTP group.
Workflows and Data Management. Workflow and DM Run3 and after: conditions m LHCb major upgrade is for Run3 (2020 horizon)! o Luminosity x 5 ( )
Hardware Status Online CRRC08 prep meeting Niko Neufeld.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Hands-On Microsoft Windows Server 2008 Chapter 5 Configuring Windows Server 2008 Printing.
Ideas and test setup for data transfer from CERN to Italian Ground Segment. M. Boschini, A. Favalli, M. Levtchenko CERN – March, 31, 2003.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Status of tests in the LCG 3D database testbed Eva Dafonte Pérez LCG Database Deployment and Persistency Workshop.
Level 1-2 Trigger Data Base development Current status and overview Myron Campbell, Alexei Varganov, Stephen Miller University of Michigan August 17, 2000.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
M.Frank, CERN/LHCb Persistency Workshop, Dec, 2004 Distributed Databases in LHCb  Main databases in LHCb Online / Offline and their clients  The cross.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
1 Copyright © 2008, Oracle. All rights reserved. Repository Basics.
Real Time Fake Analysis at PIC
Database Replication and Monitoring
IT-DB Physics Services Planning for LHC start-up
Data Management and Database Framework for the MICE Experiment
Database Corruption Advanced Recovery Techniques|
Controlling a large CPU farm using industrial tools
Grid Application Support Group Case study Schrodinger equations on the Grid Status report 16. January, Created by Akos Balasko
Event Storage GAUDI - Data access/storage Framework related issues
Status and plans for bookkeeping system and production tools
Presentation transcript:

11/01/20081 Data simulator status CCRC’08 Preparatory Meeting Radu Stoica, CERN* 11 th January 2007 * On leave from IFIN-HH

11/01/20082 How the data challenge works (Online view) A set of MDF files are produced in advance. They are copied to the Online storage in the pit. A data simulator will create symbolic links to files and register them in the Run Database. Files are copied to CASTOR and added to the LCG File Catalog. The files are added to the LHCb Bookkeeping database. Files are available for offline reconstruction in GRID. After migration, files (or rather the links to them) can be deleted in the Online storage

11/01/20083 Two possibilities for creating a file generator from the MDF files: 1. Register them directly as if created by the data writer a) Preparatory step (necessary to avoid duplication of IO): Data Simulator Online SAN MDF Files Script Database Calculates the necessary checksums Saves other file information copy read

11/01/20084 Two possibilities for creating a file generator from the MDF files: 1. Register them directly as if created by the data writer b) File generator: Data Simulator Online SAN Data Simulator Database 4. Creates runs 5. Generates file entries (name, directory structure, etc) 6. Closes runs 2. Directory structure 3. Symlinks Online Run DB 1. MDF file data Replaces ECS

11/01/20085 Two possibilities for creating a file generator from the MDF files: 1. Register them directly as if created by the data writer Pros/Cons: Easy to implement/use. No need to multiply IO to Online SAN Less software components involved (no special control system) Small number of machines required (1-2 servers) Easy to control rate (just add a new file entry at desired interval) Questions: Is the fact that the event time in the MDF files will be always the same a problem? (e.g. because of the conditions db) Data Simulator

11/01/20086 Two possibilities for creating a file generator from the MDF files: 2. Replace event builder and sent them directly to the streaming and formatting layer: Data Simulator Online SAN Data Simulator Extract events from MDF files Online Run DB replaces farm Streaming Formatting Data Writer ECS ? ?

11/01/20087 Two possibilities for creating a file generator from the MDF files: 2. Replace the farm and sent them directly to the streaming and formatting layer Pros/Cons/Observations: Requires 3x IO to storage (reading, writing, reading). At least 6 servers needed (if we want to scale up the system). More software components tested (also the ECS, Streaming/Formatting layer, Data Writer are involved). The rate can also be controlled by monitoring the writing time of a file. Data Simulator

11/01/20088 A new LUN was created on the Online SAN (30 TBs), the same size as the CASTOR RAW pool. The IO file migration requirements are: –Minimum 75 MB/s. This is the expected average rate during the data challenge. –Maximum 150 MB/s as agreed with IT. The actual IO requirements are multiplied depending on how many times a file is to be read/written. Using the 10 Gb/s network connection to CASTOR. The store03, store04 servers will be used to migrate files. Storage

11/01/20089 There is a need to create a clear picture of the whole process and decide responsibilities. Can the results of today’s discussion be written to some document? We should have a test setup as soon as possible. What is needed exactly for this? Suggestions/Questions