AMS COMPUTING VIII International Workshop Advanced Computing And Analysis Techniques in Physics Research Moscow, June 24-29, 2002 Vitali Choutko, Alexei.

Slides:



Advertisements
Similar presentations
Computing Strategy of the AMS-02 Experiment B. Shan 1 On behalf of AMS Collaboration 1 Beihang University, China CHEP 2015, Okinawa.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Building Large Scale Fabrics – A Summary Marcel Kunze, FZK.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
AMS Computing Y2001-Y2002 AMS Technical Interchange Meeting MIT Jan 22-25, 2002 Vitali Choutko, Alexei Klimentov.
CERN 14/01/20021 Data Handling Scheme for the Italian Ground Segment (IGS), as part of AMS-02 Ground Segment (P.G. Rancoita) Functions of a “Regional Center”
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
AMS TIM, CERN Apr 12, 2005 AMS Computing and Ground Centers Status Report Alexei Klimentov —
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
OO Software and Data Handling in AMS Computing in High Energy and Nuclear Physics Beijing, September 3-7, 2001 Vitali Choutko, Alexei Klimentov MIT, ETHZ.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
AMS-02 Computing and Ground Data Handling CHEP 2004 September 29, Interlaken Alexei Klimentov — ETH Zurich and
Status of AMS Regional Data Center The Second International Workshop on HEP Data Grid CHEP, KNU G. N. Kim, J. W. Shin, N. Tasneem, M. W.
VLBI Storage Manager (VSM) Matteo Stagni – Mauro Nanni 0.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
The ALICE short-term use case DataGrid WP6 Meeting Milano, 11 Dec 2000Piergiorgio Cerello 1 Physics Performance Report (PPR) production starting in Feb2001.
Systems in AMS02 AMS July 2003 Computing and Ground MIT Alexei Klimentov —
CHEP as an AMS Remote Data Center International HEP DataGrid Workshop CHEP, KNU G.N. Kim, H.B. Park, K.H. Cho, S. Ro, Y.D. Oh, D. Son (Kyungpook)
2-3 April 2001HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
RAL Site report John Gordon ITD October 1999
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Computing Strategy of the AMS-02 Experiment V. Choutko 1, B. Shan 2 A. Egorov 1, A. Eline 1 1 MIT, USA 2 Beihang University, China CHEP 2015, Okinawa.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
Computing Strategy of the AMS-02 Experiment
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
AMS02 Data Volume, Staging and Archiving Issues AMS Computing Meeting CERN April 8, 2002 Alexei Klimentov.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
AMS02 Software and Hardware Evaluation A.Eline. Outline  AMS SOC  AMS POC  AMS Gateway Computer  AMS Servers  AMS ProductionNodes  AMS Backup Solution.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Evolution of Monitoring System for AMS Science Operation Center
PC Farms & Central Data Recording
LHC experiments Requirements and Concepts ALICE
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
SAM at CCIN2P3 configuration issues
Short to middle term GRID deployment plan for LHCb
Presentation transcript:

AMS COMPUTING VIII International Workshop Advanced Computing And Analysis Techniques in Physics Research Moscow, June 24-29, 2002 Vitali Choutko, Alexei Klimentov MIT Cambridge, ETH Zurich

A.KlimentovAMS Computing ACAT'20022 Outline AMS – particle physics experiment on the international space station : AMS Data flow and ground centers Processing Farm HW components choice AMS HW evaluation Software development Distributed Data Processing STS91 precursor flight AMS ISS mission

A.KlimentovAMS Computing ACAT'20023 AMS : a particle physics experiment in space PHYSICS GOALS : Accurate, high statistics measurements of charged, cosmic ray spectra in space > 0.1GV Nuclei and e- spectra measurement The study of dark matter (90% ?) Determination of the existence or absence of antimatter in the Universe Look for negative nuclei The study of the origin and composition of cosmic rays Measure isotopes D, He, Li, Be… +

A.KlimentovAMS Computing ACAT'20024 Magnet : Nd 2 Fe 14 B TOF : trigger, velocity and Z Si Tracker : charge sign, rigidity, Z Aerogel Threshold Cerenkov : velocity Anticounters : reject multi particle events Results :  Anti-matter search : He / He = 1.1x 10  Charged Cosmic Ray spectra Pr, D, e-, He, N  Geomagnetic effects on CR under/over geomagnetic cutoff components 10 events recorded Trigger rates 0.1-1kHz DAQ lifetime 90% 8 -6 _ Precursor flight : +

A.KlimentovAMS Computing ACAT'20025 AMS Detector in shuttle payload bay (STS91 flight)

A.KlimentovAMS Computing ACAT'20026 AMS on ISS, 3 years in space Separate e- from p,p + _ up to 300 GeV He, He, B, C… 3 4 e-,   up to 1000 GeV +

A.KlimentovAMS Computing ACAT'20027

A.KlimentovAMS Computing ACAT'20028 AMS02 Data Volume Average event rate Hz (after Level-3 trigger) “Raw” Data : StreamAv.BandwidthGB/dayTB/year Scientific and Calibration 3-4 MBit/s (orbit average) House Keeping 16 Kbit/s (continuously) NASA ancillary data Event Summary Data : 45 TB/year Monte-Carlo Data : 45 TB/year The Total Data Volume of AMS PByte

A.KlimentovAMS Computing ACAT'20029 Monitoring & science data Stored data Real-time Data H&S Real-time & “Dump” data Real-time, “Dump”, & White Sand’s LOR playback AMS Crew Operation Post High Rate Frame MUX White Sand, NM facility MSFC, Al Payload Data Service system Telescience centers External Communications GSE Long Term Short Term Payload Operations Control Center Science Operations Center ISS to Remote AMS Centers Data Flow ISS NASA Ground Infrastructure Remote AMS Sites H&S Monitoring Science Flight ancillary data Real-time & “dump” NearReal-time File transfer playback

A.KlimentovAMS Computing ACAT' AMS Ground Centers Science Operations Center POCC AL AMS Remote center RT data Commanding Monitoring NRT Analysis NRT Data Processing Primary storage Archiving Distribution Science Analysis MC production Data mirror archiving External Communications Science Operations Center XTerm HOSC Web Server and xterm TReK WS commands Monitoring, H&S data Flight Ancillary data AMS science data (selected) TReK WS “ voice”loop Video distribution Production Farm Analysis Facilities PC Farm Data Server Analysis Facilities GSE D S A e T r A v e r GSE Buffer data Retransmit To SOC AMS Station AMS Station AMS Station GSE MC production cmds archive AMS Data, NASA data, metadata

A.KlimentovAMS Computing ACAT' AMS Ground Centers (Summary) Ground Support Computers at MSFC Huntsville Al  receives data from NASA  buffers until retransmission to SOC  runs unattended 24/day, 7 days/week Payload Operations and Control Center  single commanding point  monitor data and run quality control program  process about 10% of data in near real-time mode Science Operations Center (CERN)  Data production and MC simulation  Reconstructs 100% of events, typical delay less than 1 day  Central analysis system provides most of required analysis power  Complete data and MC repositary Remote Centers (Italy, USA)  Mirror data and metadata repositary  Monte Carlo production  Analysis facility to support local user’s community

A.KlimentovAMS Computing ACAT' AMS02 HW Choice Processing Farm (poor man solution) To build high-performance quality computing farm with cheap computing elements  The choice of the operating system is easy (Linux is de facto standard for the farms)  Networking components are also well identified (Gigabit switch)  CPU and Disks nodes processor and server nodes have the identical architecture (dual-CPU, 1GB RAM, 3 ethernet interfaces), computers equipped with IDE RAID arrays, servers have IDE and SCSI RAID arrays (data processing is done on both)

A.KlimentovAMS Computing ACAT' AMS02 HW Choice (cont’d) AMS02 Farm Prototype :  processors : 8 nodes (PIII, AMD, PIV) equipped with 3-ware escalade controller and TB RAID5 array of IDE disks. Events processing is done localy and the results are stored on “local disk”. 3 production jobs per machine.  servers : 2 nodes (AMD and PIV Xeon), each node has 0.8 TB RAID5 array of IDE disks and in addition external UW SCSI tower of 1.4TB, nfs’ed via dedicated ethernet segment to processing nodes. 2 production jobs per machine. We couldn’t reach declared 160 MB/s for UW SCSI controller, but 80 MB/sec). Data moved from processor nodes to server on run-by-run basis (6 months of extensive disks usage, SCSI RAID I/O performance is 20% better) 2 USD/Gbyte for “IDE RAID” and 10 USD/Gbyte for “SCSI RAID”

A.KlimentovAMS Computing ACAT' Evaluation of CPU performance Use AMS Offline simuation/reconstruction code compiled with CERNlibs, Geant, Naglib, PAW, ROOT, etc

A.KlimentovAMS Computing ACAT' AMS02 Benchmarks Executive time of AMS “standard” job compare to CPU clock 1) V.Choutko, A.Klimentov AMS note ) Brand, CPU, Memory Intel PII dual-CPU 450 MHz, 512 MB RAM OS/Compiler RH Linux 6.2 / gcc 2.95 “Sim” 1 “Rec” 1 Intel PIII dual-CPU 933 MHz, 512 MB RAMRH Linux 6.2 / gcc Compaq, Quad α-ev MHz, 2 GB RAMRH Linux 6.2 / gcc AMD Athlon,1.2GHz, 256 MB RAMRH Linux 6.2 / gcc Intel Pentium IV 1.5GHz, 256 MB RAMRH Linux 6.2 / gcc Compaq dual-CPU PIV Xeon 1.7GHz, 2GB RAMRH Linux 6.2 / gcc Compaq dual α-ev68 866MHz, 2GB RAMTru64 Unix/ cxx Elonex Intel dual-CPU PIV Xeon 2GHz, 1GB RAMRH Linux 7.2 / gcc AMD Athlon 1800MP, dual-CPU 1.53GHz, 1GB RAM RH Linux 7.2 / gcc CPU SUN-Fire-880, 750MHz, 8GB RAMSolaris 5.8/C CPU Sun Ultrasparc-III+, 900MHz, 96GB RAMRH Linux 6.2 / gcc Compaq α-ev68 dual 866MHz, 2GB RAMRH Linux 7.1 / gcc

A.KlimentovAMS Computing ACAT'200216

A.KlimentovAMS Computing ACAT'200217

A.KlimentovAMS Computing ACAT' AMS02 Benchmarks (summary) α-ev68 866MHz and AMD Athlon MP have nearly the same performance and are the best candidates for “AMS processing node” (the price of system based on α-ev68 is twice higher than the similar one based on AMD Athlon) Though PIV Xeon has lower performance, resulting 15% overhead comparing with AMD Athlon MP 1800+, the requirements of high reliability for “AMS server node” dictates the choice of Pentium machine. SUN and COMPAQ SMP might be the candidates for AMS analysis computer (the choice is postponed up to the fall of 2004)

A.KlimentovAMS Computing ACAT' AMS Software Strategies, choices and basic decisions :  new code C++ only (though we had a large part of legacy SW written on Fortran)  Existing Fortran libraries (Cernlib, Geant, Naglib, etc) incorporated via C/Fortran interface (R.Burow)  Simualation code compatible with both “Geants”  Geant3 vs Geant4, Root vs Ntuples are selected by Makefile options  transient and persistent classes are separated with implementing of copy member functions  Use Root and HBOOK for histogramming and data visualization

A.KlimentovAMS Computing ACAT' AMS Software (cont’d)  Use different persistency solutions for various type of data : Flat files for “static”data like geometry and magnetic field map Flat files for raw data Ntuples and Root files for ESD Relational Database (Oracle) [Objectivity up to Sep 1998] o Calibration data o Event Tags o Slow control data o NASA ancillary data o Full processing history o Input paramaters o Computing and storage facilities o Files location catalogues  Keep the data store and data catalogues separate

A.KlimentovAMS Computing ACAT' AMS Software (cont’d)  TReK for commanding and “online” monitoring (Telescience Resource Kit, NASA certified SW)  AMS adopted BBFTP version for data transferring between MSFC Al and CERN ( bbftp was designed in BaBar to transfer big files between SLAC and Lyon)  AMS data processing system based on Corba client- server model and Oracle RDBMS  Distributed MC Data Production System

A.KlimentovAMS Computing ACAT' server producers Raw data Oracle RDBMS Conditions DB Tag DB Active Tables : Hosts, Interfaces, Producers, Servers Catalogues Nominal Tables Hosts, Interfaces Producers, Servers… ESD Raw data {I} {II} {III} {IV} {V} {VI} {I} submit 1 st server {II} “cold” start {III} read “active” tables (available hosts, number of servers, producers, jobs/host) {IV} submit servers {V} get “run”info (runs to be processed, ESD output path) {VI} submit producers (LILO, LIRO,RIRO…) Notify servers AMS Production server

A.KlimentovAMS Computing ACAT' AMS Production Highlights Stable running for more than 9 months Average efficiency 95% (98% without Oracle) Processes communication and control via Corba LSF and remote shell for process submission Run Oracle server on AS4100 Alpha and Oracle clients on Linux. Oracle RDBMS (for what it is good)  Tag DB  Conditions DB  Bookkeeping  Production status  Runs history  Files location catalogues

A.KlimentovAMS Computing ACAT' Distributed MC Production Central repositary with binaries, execs, input parameters Static version of all exec files Central Relational Database to keep description of MC centers and computing facilities MC users Production jobs status and processing history Files location catalogues Production status  Central server (use CORBA) to control local and remote clients  Central Database Server (initially C++, now Perl)  WEB interface to control and to monitor production status to download binaries, geometry, magnetic field map, etc

A.KlimentovAMS Computing ACAT' Conclusions AMS02 production farm prototype concept is successfully tested. The prototype is in use. Data production SW has proven to be stable and to deliver 95% performance for the period of weeks. MC distrubuted production system is under testing with the remote sites in Europe (France, Italy, Suisse) and US (MIT)