PC Farms & Central Data Recording

Slides:



Advertisements
Similar presentations
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Advertisements

Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/20 New Experiences with the ALICE High Level Trigger Data Transport.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
14th April 1999Hepix Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
CERN - European Laboratory for Particle Physics DESY November 2, 1998 PC Farms at CERN Frédéric Hemmer CERN-IT/PDP.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
The ALICE DAQ: Current Status and Future Challenges P. VANDE VYVRE CERN-EP/AID.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
Spending Plans and Schedule Jae Yu July 26, 2002.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
22nd March 2000HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
2-3 April 2001HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
HPSS for Archival Storage Tom Sherwin Storage Group Leader, SDSC
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
4-8 th October 1999CERN Site Report, HEPiX SLAC. A.Silverman CERN Site Report HEPNT/HEPiX October 1999 SLAC Alan Silverman CERN/IT/DIS.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
RAL Site report John Gordon ITD October 1999
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
HepNT - January 15, 1997 : PCSF Frederic Hemmer IT/PDP 1 PCSF - A Pentium ® /Windows NT ® Based simulation farm Frederic Hemmer CERN IT/PDP.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
14 th April 1999CERN Site Report, HEPiX RAL. A.Silverman CERN Site Report HEPiX April 1999 RAL Alan Silverman CERN/IT/DIS.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Linux IDE Disk Servers Andrew Sansum 8 March 2000.
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
Virtual Server Server Self Service Center (S3C) JI July.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Gu Minhao, DAQ group Experimental Center of IHEP February 2011
ALICE Computing Data Challenge VI
LHC experiments Requirements and Concepts ALICE
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
SAM at CCIN2P3 configuration issues
Online Software Status
Grid Canada Testbed using HEP applications
Example of DAQ Trigger issues for the SoLID experiment
CASTOR: CERN’s data management system
The Performance and Scalability of the back-end DAQ sub-system
Lee Lueking D0RACE January 17, 2002
Cluster Computers.
Presentation transcript:

PC Farms & Central Data Recording CERN - European Laboratory for Particle Physics ATLAS Trigger/DAQ Workshop Chamonix, October 20, 1998 Frédéric HEMMER - CERN/IT Http://wwwinfo.cern.ch/pdp/pc/EF/ATLAS-TDAQ-1098

Frédéric Hemmer CERN/IT Overview NA48 Data Recording NA45 Data Recording in Objectivity NA57 Data Recording in HPSS Summary ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

NA48 Central Data Recording Sub detector VME crates Cisco 5505 Event Builder Online PC Farm FDDI Fast Ethernet SUN E450 500 GB Disk space XLNT Gbit 3Com 3900 Fast Ethernet 7 KM Gigabit Ethernet 3Com 9300 GigaRouter HiPPI HiPPI FDDI Offline PC Farm CS/2 2.5 TB Disk space ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT NA 48 Data Recording in 98 May  September 1998 Raw Data on Tape 68 TB (1450 tapes, mainly 50 GB tapes) 12.5 TB Selected Reconstructed Data Total with 97 data : 96 TB Average Data Rate : 18 MB/s (peaks @ 23 MB/s) CDR system can do 40-50 MB/s; limitation is CPU Time available Data recorded as files (4 million) ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT NA48 On Line Farm 11 Subdetector PC’s (dual PII-266, 128 MB) 8 Event Building PC’s (dual PII-266, 128 MB, 18 GB SCSI) 4 CDR routing PC’s (dual PII-266, 64 MB, FDDI) All running Linux Software event building in the interburst gap Optional Software Filter (tags data) Send data to computer center (local disk buffers : 144 GB , 2 hours) On CS/2 : L3 Filtering and tape writing ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT NA48 Plans for 1999 Sub detector VME crates Fast Ethernet Cisco 5505 Event Builder 7 KM 4 * SUN E450 4.5 TB Disk space Gigabit Ethernet Fast Ethernet 3Com 3900 3Com 9300 HiPPI HiPPI Gigabit Ethernet On/Offline PC Farm ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Sub detector VME crates Frédéric Hemmer CERN/IT NA45 Data Recording Sub detector VME crates NA48 SCI Event Builder On Line PC Farm Fast Ethernet 3Com 3900 PCSF 7 KM Gigabit Ethernet 2 * SUN E450 500 GB Disk space Fast Ethernet 3Com 9300 HiPPI HiPPI 3Com 3900 Gigabit Ethernet ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

NA45 Raw Data recording in Objectivity October 98 ; November 98 Estimated bandwidth : 15 MB/s Processes translate Raw Data format to Objectivity Database files (1.5 GB) are closed, then written on tape Steering done using a set of perl scripts on the disk servers On line filtering/reconstruction/calibration possible Farm is running Windows NT Reconstruction can use PCSF ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT PCSF Configuration (1) Server running NT 4.0 Server SP3 1 dual capable Ppro @ 200 MHz, 96 MB, with 9 GB data disk (with mirroring). LSF central queues. Server running NT Terminal Server Beta 2 1 dual Ppro @ 200 MHz, 128 MB, with 4 GB data disk. Runs IIS 3.0 and is accessible from outside CERN. It also host the asp’s for Web access Servers running NT 4.0 Workstation SP3 9 dual Ppro’s @ 200 MHz, 64 MB, 2*4GB 25 dual PII’s @ 300 MHz, 128 MB, 2*4GB All equipped with boot proms We requested 2 GB disks, they were unable to deliver given the time between the quote and the order. They gave 4 GB @ 35$ increase in price per disk. ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT PCSF Configuration (2) Machines interconnected with 4 3com 3000 100BaseT switch Display/Keyboard/Mouse connected to a Raritan multiplexor PC Duo for remote admin access  There were problems with other products All running LSF 3.0.  LSF 3.2 does not work, support weak Completely integrated with NICE 4*3000 -> 2*3900 with Gbit Enet Some Ethernet cards did not work on some PC’s The same card marketing name may come with different chip set (e.g. 82557, 82558, 82559) PC Anywhere gave problems on MP machines Remotely possible did not install well on unattended setup ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT Racking evolution 1997 1998 ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT HPSS test system A schematic view of the CERN HPSS test installation. Shown is also the NA57 CDR computer. As can see the NA57 CDR to HPSS is over FDDI. The CERN HPSS test system consists of: one IBM G40 as HPSS main server and disk and 3590 tape mover (very busy machine) one IBM F50 (1 CPU) as disk mover with SSA disks one DEC Alpha 500 with disk mover and Redwood tape mover and STK PVR. This node was also used for the DEC port. ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

HPSS Tests results (NA57) Disk to disk transfers without tape migration: sustained: 5 - 6MB/s (1GB files) 4-5MB/s (200MB files) average: ~5.5MB/s peak: 6.5MB/s With disk to tape migration: sustained: 2 - 3MB/s (200MB files) average: 3 - 5MB/s peak: 5.5 - 6.5MB/s Disk @ DAQ limited to 7 MB/s read ! Pure disk to disk transfers with 1GB files gave good rate, 5-6MB/s sustained whereas with 200MB files we only achieved 4-5MB/s. This difference is explained by delays in our own CDR software: as it was configured it took in average about 5 seconds for it to detect a new file and for a 200MB file this is about 20% of the total transfer time. With migration the bandwidth into the HPSS main server, the G40, put additional constraints on the transfer rates. The conclusion from the tests with our HPSS test system was that it was not suitable for performance tests. However we learned where our bottlenecks are, e.g. local migration, dead-times in the CDR software. HPSS itself does not seem to have put any constraint. It used the H/W as it could. ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT Production system MPPC2604 Barracuda-9, 18GB 3900 switch Gbit Ethernet STK Powderhorn silos Redwood 100BaseT Computer center 3590 NA57 FDDI HiPPI ~120 GB mirrored ~120 GB mirrored IBM 3494 robot RS6000/F50, 2CPU, 512MB 2 * DEC Alpha 4100 4 CPU, 512Mb RS6000/F50, 2CPU, 256MB During this (‘98) summer an HPSS production system has been setup at CERN. It consists of: two IBM F50 with 2 CPUs and 512 and 256 MB/s memory one (or two) DEC (Compaq) Alpha 4100 with 4 CPUs and 512MB memory Two disk storage classes with 120GB mirrored disk each one 3590 tape storage class with two drives one Redwood tape (50GB) storage class with 2 or 3 drives. Two pre-production services are about to start: NA57 CDR to the DEC disk storage class in hierarchy with the Redwoods. “user tapes” on the IBM disk storage class in hierarchy with the 3590s. First preliminary tests of transfers between NA57 and the DEC disk storage class gave between 5.5 and 7MB/s over FDDI. If ready in time a new GBit ethernet link will be used, which should allow about 10MB/s band\width. ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Current & Future Data rates ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT

Frédéric Hemmer CERN/IT Summary On line PC farms are being used to record data at sensible rates (Linux) Off line PC farms are being used for reconstruction/filtering/analysis (Linux/NT) New paradigms of recording data are explored (Objectivity/HPSS) Still a lot to do on scalable farm management, global steering, CDR monitoring, etc.. ATLAS T/DAQ Workshop-Chamonix 20 Oct. 98 Frédéric Hemmer CERN/IT