STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.

Slides:



Advertisements
Similar presentations
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Advertisements

June 19, 2002 A Software Skeleton for the Full Front-End Crate Test at BNL Goal: to provide a working data acquisition (DAQ) system for the coming full.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.
ATLAS Analysis Model. Introduction On Feb 11, 2008 the Analysis Model Forum published a report (D. Costanzo, I. Hinchliffe, S. Menke, ATL- GEN-INT )
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
STAR Software Basics Introduction to the working environment Lee Barnby - Kent State University.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
ROOT An object oriented HEP analysis framework.. Computing in Physics Physics = experimental science =>Experiments (e.g. at CERN) Planning phase Physics.
GRID job tracking and monitoring Dmitry Rogozin Laboratory of Particle Physics, JINR 07/08/ /09/2006.
CFT Offline Monitoring Michael Friedman. Contents Procedure  About the executable  Notes on how to run Results  What output there is and how to access.
SVX Software Overview Sasha Lebedev VTX meeting 09/07/ SVX Software web page:
Introduction to Hall-D Software February 27, 2009 David Lawrence - JLab.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Current Status of Hadron Analysis Introduction Hadron PID by PHENIX-TOF  Current status of charged hadron PID  CGL and track projection point on TOF.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
9/26/11HFT soft meeting, BNL1 Chain analysis fz file MuDst.root minimc.root geant.root event.root McEvent.root StMiniMcMaker StAssociationMaker : STAR.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
HERA/LHC Workshop, MC Tools working group, HzTool, JetWeb and CEDAR Tools for validating and tuning MC models Ben Waugh, UCL Workshop on.
Guide to Linux Installation and Administration, 2e1 Chapter 10 Managing System Resources.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
STAR Analysis Meeting, BNL, Dec 2004 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software and calibration L3 display 200 GeV February.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Computer Science Research and Development Department Computing Sciences Directorate, L B N L 1 Storage Management and Data Mining in High Energy Physics.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Andreas Morsch, CERN EP/AIP CHEP 2003 Simulation in ALICE Andreas Morsch For the ALICE Offline Project 2003 Conference for Computing in High Energy and.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
Alexander Richards, UCL 1 Atlfast and RTT (plus DCube) Christmas Meeting 18/12/2007.
October 8, 2002P. Nilsson, SPD General Meeting1 Paul Nilsson, SPD General Meeting, Oct. 8, 2002 New tools and software updates Test beam analysis Software.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
CEBAF The Continuous Electron Beam Accelerating Facility(CEBAF) is the central particle accelerator at JLab. CEBAF is capable of producing electron beams.
February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.
1D. Olson, SDM-ISIC Mtg, 26 Mar 2002 Scientific Data Management: An Incomplete Experimental HENP Perspective D. Olson, LBNL 26 March 2002 SDM-ISIC Meeting.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
PHENIX Simulation System 1 January 12, 2000 Simulation: Status for VRDC Tarun Ghosh, Indrani Ojha, Charles Vanderbilt University.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
STAR C OMPUTING Plans for Production Use of Grand Challenge Software in STAR Torre Wenaus BNL Grand Challenge Meeting LBNL 10/23/98.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Experience with CalcHEP H. S. Goh Univ. of Arizona very little West Coast LHC Theory Network -- UC Irvine May
2007 Run Update for STAR Jeff Landgraf For the STAR collaboration.
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
1 Checks on SDD Data Piergiorgio Cerello, Francesco Prino, Melinda Siciliano.
STAR J/  Trigger in dA Manuel Calderon for the Heavy-Flavor Group Trigger Workshop at BNL October 21, 2002.
 Distributed Database Concepts  Parallel Vs Distributed Technology  Advantages  Additional Functions  Distribution Database Design  Data Fragmentation.
Analyzing ever growing datasets in PHENIX Chris Pinkenburg for the PHENIX collaboration.
M. Oldenburg GridPP Metadata Workshop — July 4–7 2006, Oxford University 1 Markus Oldenburg GridPP Metadata Workshop July 4–7 2006, Oxford University ALICE.
October Test Beam DAQ. Framework sketch Only DAQs subprograms works during spills Each subprogram produces an output each spill Each dependant subprogram.
TAGS in the Analysis Model Jack Cranshaw, Argonne National Lab September 10, 2009.
Overview of PHENIX Muon Tracker Data Analysis PHENIX Muon Tracker Muon Tracker Software Muon Tracker Database Muon Event Display Performance Muon Reconstruction.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
STAR Simulation. Status and plans V. Perevoztchikov Brookhaven National Laboratory,USA.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
BES III Software: Beta Release Plan Weidong Li 19 th October 2005.
Object-Oriented Track Reconstruction in the PHENIX Detector at RHIC Outline The PHENIX Detector Tracking in PHENIX Overview Algorithms Object-Oriented.
Quark Matter 2002, July 18-24, Nantes, France Dimuon Production from Au-Au Collisions at Ming Xiong Liu Los Alamos National Laboratory (for the PHENIX.
Tracking, Computing & other Stuff. Correlation of detector hits The track segments of inner and outer MDCs are matched on Cluster level The track segments.
Review and Perspective of Muon Studies Marco Destefanis Università degli Studi di Torino PANDA Collaboration Meeting GSI (Germany) September 5-9, 2011.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL May 19, 2003 BNL Technology Meeting.
ANALYSIS TRAIN ON THE GRID Mihaela Gheata. AOD production train ◦ AOD production will be organized in a ‘train’ of tasks ◦ To maximize efficiency of full.
HYDRA Framework. Setup of software environment Setup of software environment Using the documentation Using the documentation How to compile a program.
Data Analysis in Particle Physics
Presentation transcript:

STAR Software Walk-Through

Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data to look at! In 2007, ~70M minimum bias events. – Need computers to be able to analyze vast dataset. Analysis Software: – Find relevant data – Develop analysis code Interface to the data Analysis code: plots and correlations of experimentally measured quantities – Submitting jobs to the batch farm – Plotting results

Computing resources Facilities available for STAR – PDSF (in Berkeley, part of NERSC) Parallel Distributed Systems Facility – RCF (in Brookhaven Lab) RHIC Computing Facility Log in to PDSF: pdsf.nersc.gov – Note for PDSF: make sure you have a.chos file. I selected /auto/redhat8 as my environment. – Without an environment properly set, perl won’t work! Web page: – FAQ – USER accounts – Monitoring of farm conditions – … Log in to RCF: rssh.rhic.bnl.gov Web page:

Doing Analysis 101 For real or simulated data that has already been produced into a standard format: Find Data HPSS NFS local. Tools: FileCatalog Run analysis on data. Tools: MuDST StMcEvent Plot results of analysis. Tools: ROOT classes.

Tools: FileCatalog (get_file_list.pl) – Finding Files (of course) that satisfy certain conditions: – production library used, trigger setup used, run numbers, collision system, … i.e. it is a database of our data. Examples of usage found in link above. Production Location – STAR → Computing → Production Location –

Tools: Scheduler – Used to submit jobs to the RCAS linux farm in batch mode. – RCAS: ~10 interactive nodes, ~150 batch nodes. – How to use it: XML script that specifies – files to be used (e.g. using a catalog query) – macro to be executed (i.e. analysis to be done) – what to do with the output – STAR → Computing → Scheduler

Analyzing Example: Real Data, Step I Find events of interest – Au Au collisions, 200 GeV. – p+p collisions, 200 GeV Many collisions and triggers have been used. Example: Looking at “Minimum bias” triggers – This trigger is meant to capture almost all interactions. – Every trigger detector introduces a bias, this trigger is meant to reduce the bias introduced as much as possible. Trigger ID: – Each file can have events that were selected by various trigger conditions – ID picks out a given trigger condition. – 2004 list of triggers: – 2009 data:

A file catalog query… output: path and filename of file. query conditions: – Production: P10ic, production from 2010, real data, official production (pp2009 data) – Filetype: daq_reco_mudst: came from DAQ (real data), processed through the reconstruction chain, stored into a micro Data Summary Tape. – storage: NFS. Mounted on hard disks accessible through the Networked File System, as opposed to the files stored on tape drives in HPSS. ~/afsdir/wrk/jpsi/offline/> get_file_list.pl -keys 'path,filename' -cond 'production=P10ic,filetype=daq_reco_mudst,storage=NFS,sanity=1,filename~st _upsilon' -limit 10 /star/data81/reco/production2009_200Gev_Single/ReversedFullField/P10ic/2009/1 38/ ::st_upsilon_adc_ _raw_ MuDst.root /star/data76/reco/production2009_200Gev_Single/ReversedFullField/P10ic/2009/1 39/ ::st_upsilon_adc_ _raw_ MuDst.root /star/data81/reco/production2009_200Gev_Single/ReversedFullField/P10ic/2009/1 38/ ::st_upsilon_adc_ _raw_ MuDst.root

What’s in a MuDST? The result of the reconstruction of an event. – Trigger information. Signal in ZDC, BBC, VPD, EMC, … – Track information. Number of points found by tracker and used in fit Momentum at first point and last point (and helix parameters) Covariance matrix of track fit dE/dx – nSigma, PID probability charge id of track in event (unique number for a track in an event) type (0=global, 1=primary)  2 Topology Map: bit pattern of hits in detector Distance of closest approach to vertex (for global) Position of first and last points TBrowser, check contents of file. ROOT objects can be drawn quickly.

Define an Analysis Task Examples: – Multiplicity Distribution ~Probability to find events with Nch tracks. Nch: number of charged particles in the event (typically, per unit rapidity at midrapidity). – pT distribution of charged tracks for all events. ~Probability to find a track with a given pT.

The “Maker” framework “Makers” are a way to standardize the way we do analysis: – All have to “prepare” or initialize e.g. book histograms and trees – All do something every event e.g. calculate or obtain distributions of interest – All have to clean up when the job is done e.g. write the histograms

Example Maker code: In PDSF: /auto/pdsfdv39/starspec/pdsfdv34/starspec/calderon/tutorial s/StRoot/StMuDstExampleAnalysisMaker All Makers live under a directory called StRoot Compilation of the analysis code is done in same directory where StRoot directory (or link) is found – cons +StMuDstExample Running is done in same directory where compilation was done. Example in StRoot/macros/examples/ – root4star –b –q ‘RunMuDstExample.C(500)'

Plotting the results Open the output file (dummyFile00.root) in root. Can issue C++ commands on the root classes interactively. Set a sensible color scheme – gStyle->SetPalette(1,0); Create canvases (TCanvas) – TCanvas* cnv1 = new TCanvas(“cnv1”,”multiplicity”,600,600); For drawing histograms: – TH1::Draw() – mMult->Draw(); Can change line color, width, style – mMult->SetLineColor(2); – mMult->SetLineWidth(3); – mMult->SetLineStyle(11); Can draw markers – mMult->SetMarkerStyle(20); – mMult->Draw(“P”); For reproducibility, can also put all the commands into a macro, and just execute the macro: –.x plotMultExample.C

Analyzing Example: Simulation, Step I Generate events of interest – e.g. Lambda_c, Upsilon, J/Psi particles according to a distribution – e.g. use event generators PYTHIA HERWIG HIJING NEXUS AMPT MPC For large datasets, request is done officially to STAR simulation (Maxim Potekhin, simulation leader)