PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,

Slides:



Advertisements
Similar presentations
1 GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
Advertisements

Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Terabyte IDE RAID-5 Disk Arrays David A. Sanders, Lucien M. Cremaldi, Vance Eschenburg, Romulus Godang, Christopher N. Lawrence, Chris Riley, and Donald.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
CCJ Computing Center in Japan for spin physics at RHIC T. Ichihara, Y. Watanabe, S. Yokkaichi, O. Jinnouchi, N. Saito, H. En’yo, M. Ishihara,Y.Goto (1),
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
Design & Management of the JLAB Farms Ian Bird, Jefferson Lab May 24, 2001 FNAL LCCWS.
RIKEN CCJ Project Regional computing center in Japan for BNL-RHIC experiment especially for PHENIX collaboration. CCJ serves for RHIC physics activity.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
JPS Matsue1 PHENIX Computing Center in Japan (PHENIX CC-J) の採用技術 澤田真也( KEK ) 市原卓、渡邊康(理研、理研 BNL 研究センター) 後藤雄二、竹谷篤、林直樹(理研) 延與秀人、四日市悟(京大)、浜垣秀樹(東大.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
W.A.Wojcik/CCIN2P3, May Running the multi-platform, multi-experiment cluster at CCIN2P3 Wojciech A. Wojcik IN2P3 Computing Center
Shigeki Misawa RHIC Computing Facility Brookhaven National Laboratory Facility Evolution.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
STAR Off-line Computing Capabilities at LBNL/NERSC Doug Olson, LBNL STAR Collaboration Meeting 2 August 1999, BNL.
8 October 1999 BaBar Storage at CCIN2P3 p. 1 Rolf Rumler BaBar Storage at Lyon HEPIX and Mass Storage SLAC, California, U.S.A. 8 October 1999 Rolf Rumler,
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Sep 02 IPP Canada Remote Computing Plans Pekka K. Sinervo Department of Physics University of Toronto 4 Sep IPP Overview 2 Local Computing 3 Network.
CC-J Monthly Report Shin’ya Sawada (KEK) for CC-J Working Group
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
Nov. 8, 2000RIKEN CC-J RIKEN CC-J (PHENIX Computing Center in Japan) Report N.Hayashi / RIKEN November 8, 2000 PHENIX Computing
Disk Farms at Jefferson Lab Bryan Hess
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
National HEP Data Grid Project in Korea Kihyeon Cho Center for High Energy Physics (CHEP) Kyungpook National University CDF CAF & Grid Meeting July 12,
RAL Site report John Gordon ITD October 1999
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
RAL Site Report John Gordon HEPiX/HEPNT Catania 17th April 2002.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
November 10, 1999PHENIX CC-J Updates in Nov.991 PHENIX CC-J Updates in Nov New Hardware - N.Hayashi / RIKEN November 10, 1999 PHENIX Computing Meeting.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
May 10, 2000PHENIX CC-J Updates1 PHENIX CC-J Updates - Preparation For Opening - N.Hayashi / RIKEN May 10, 2000 PHENIX Computing
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
CCJ introduction RIKEN Nishina Center Kohei Shoji.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
PC Farms & Central Data Recording
A data Grid test-bed environment in Gigabit WAN with HPSS in Japan
RHIC Computing Facility Processing Systems
GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
Presentation transcript:

PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova, Italy

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Contents 1. Overview 2. Concept of the system 3. System Requirement 4. Other requirement as a Regional Computing Center 5. Plan and current status 6. WG for constructing the CC-J (CC-J WG) 7. Current configuration of the CC-J 8. Photographs of the CC-J 9. Linux CPU farm 10. Linux NFS performance v.s. kernel 11. HPSS current configuration 12. HPSS performance test 13. WAN performance test 14. Summary

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) PHENIX CC-J : Overview  PHENIX Regional Computing Center in Japan (CC-J) at RIKEN  Scope  Principal site of computing for PHENIX simulation  PHENIX CC-J is aiming at covering most of the simulation tasks of the whole PHENIX experiments  Regional Asian computing center  Center for the analysis of RHIC spin physics  Architecture  Essentially follow the architecture of RHIC Computing Facility (RCF) at BNL  Construction  R&D for the CC-J started in April ‘98 at RBRC  Construction began in April ‘99 over a three years period  1/3 scale of of the CC-J will be operational in April 2000

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Concept of the CC-J System

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) System Requirement for the CC-J  Annual Data amount DST150 TB micro-DST 45 TB Simulated Data 30 TB Total 225 TB  Hierarchical Storage System Handle data amount of 225TB/year Total I/O bandwidth: 112 MB/s HPSS system  Disk storage system 15 TB capacity All RAID system I/O bandwidth: 520 MB/s  CPU ( SPECint95) Simulation 8200 Sim. Reconst 1300 Sim. ana. 170 Theor. Mode 800 Data Analysis 1000 Total  Data Duplication Facility Export/import DST, simulated data.

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Other Requirements as a Regional Computing Center s Software Environment Software environment of the CC-J should be compatible to the PHENIX Offline Software environment at the RHIC Computing Facility (RCF) at BNL AFS accessibility (/afs/rhic) Objectivity/DB accessibility (replication to be tested soon) s Data Accessibility Need exchange data of 225 TB/year to RCF Most part of the data exchange will be done by SD3 tape cartridges (50GB/volume) Some part of the data exchange will be done over the WAN CC-J will use Asia-Pacific Advanced Network ( APAN ) for US-Japan connection APAN has currently 70 Mbps bandwidth for Japan-US connection Expecting 10-30% of the APAN bandwidth (7-21 M bps) can be used for this project: GB/day ( TB/year) will be transferred over the WAN

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Plan and current status of the CC-J

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Working Group for the CC-J construction (CC-J WG)  CC-J WG is a main body to construct the CC-J Hold bi-weekly regular meeting at RIKEN Wako, to discuss technical items and project plans etc. n Mailing list of the CC-J WG created (mail traffic: 1600 mails /year)

Current configuration of the CC-J

Photographs of the PHENIX CC-J at RIKEN

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Linux CPU farms s Memory Requirement : MB/CPU for a simulation chain s Node specification Motherboard: ASUS p2b Dual CPU /node (currently total 64 CPU) PentiumII (450MHz) 32 CPU + Pentium III (600 MHz) 32 CPU 512 MB memory / node (1GB SWAP/node) 14 GB HD /node (system 4GB, work 10 GB) 100 BaseT Ethernet interface (DECchip Tulip) Linux Redhat 5.2 (kernel nfsv3 patch) Portable Batch System (PBS V2.1) for batch queuing AFS is accessed through the NFS (No AFS client is installed on Linux pc) Daily mirroring of the /afs/rhic contents to a local disk file system is carrying out s PC Assemble (Alta cluster) Remote hardware-reset/power control, Remote CPU temp. monitor Serial port login from the next node (minicom) for maintenance (fsck etc.)

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Linux NFS performance v.s. kernel s NFS Performance test using bonnie benchmark for 2 GB file NFS Server : SUN Enterprise 450 (Solaris 2.6) 4 CPU (400MHz) 1GB memory NFS client : Linux RH5.2, Dual Pentium II 600 MB, 512 MB memory _ NFS performance of the recent Linux kernel seems to be improved _ nfsv3 patch is still useful for the recent kernel (2.2.14) – currently we are using the kernel nfsv3 patch – nfsv3 patch is available from

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Current HPSS hardware configuration IBM RS6000-SP 5-node (silver node: Quadruple PowerPC604e 332 MHz CPU/node ) Core server : 1, Disk mover : 2, Tape mover : 2 SP switch (300 MB/s) and 1000BaseSX NIC (OEM of Alteon) A StorageTek Powderhorn Tape Robot 4 Redwood drives and 2000 SD3 cartridges (100 TB) dedicated for HPSS Sharing the robot with other HSM systems 6 drives and 3000 cartridges for other HSM systems Gigabit Ethernet Alteon ACE180 switch for Jumbo Frame ( 9 kB MTU) Use of the Jumbo Frame reduces the CPU utilization for transfer CISCO Catalyst 2948G for distribution to 100BaseT Cache Disk : 700 GB (total), 5 components 3 SSA loops (50 GB each) 2 FW-SCSI RAID (270 GB each)

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Performance test of parallel ftp (pftp) of HPSS s pput from SUN-E450 : 12 MB/s for one pftp connection Gigabit Ethernet, Jumbo Frame (9 kB MTU) s pput from LINUX : 6 MB/s for one pftp connection 100BaseT - G.Ether - Jumbo (defragment on a switch) s Totally 〜 50 MB/s pftp performance was obtained for pput

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) WAN performance test s RIKEN (12 Mbps) - IMnet - APAN (70 Mbps) -startap- ESnet - BNL Round Trip Time for RIKEN-BNL :170 ms File transfer rate is 47 kB/s for 8 kB TCP widowsize (Solaris default) Large TCP-window size is necessary to obtain high-transfer rate RFC1323 (TCP Extensions for high performance, May 1992) describes the method of using large TCP window-size (> 64 KB)  Large ftp performance (641 kB/s = 5 Mbps) was obtained for a single ftp connection using a large TCP window-size (512 kB) over the pacific ocean (RTT = 170 ms)

Takashi Ichihara (RIKEN / RIKEN BNL Research Center) Summary s The construction of the PHENIX Computing Center in Japan (CC-J) at RIKEN Wako campus, which will extend over a three years period, began in April s The CC-J is intended as the principal site of computing for PHENIX simulation, a regional PHENIX Asian computing center, and a center for the analysis of RHIC spin Physics. s The CC-J will handle the data of about 220 TB/year and the total CPU performance is planned to be 10,000 SPECint95 in s CPU farm of 64 processors (RH5.2, kernel with nfsv3 patch) is stable. s About 50 MB/s pftp performance was obtained for HPSS access. s Large ftp performance (641 KB/s = 5 Mbps) was obtained for a single ftp connection using a large TCP window-size (512 kB) over the Pacific Ocean (RTT = 170 ms) s Stress tests for the entire system were carried out successfully. s Replication of the Objectivity/DB over the WAN will be tested soon. s The CC-J operation will be started in April 2000.