Randy MelenApril 14, 19991 Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Scale-out Central Store. Conventional Storage Verses Scale Out Clustered Storage Conventional Storage Scale Out Clustered Storage Faster……………………………………………….
NPACI Panel on Clusters David E. Culler Computer Science Division University of California, Berkeley
SLAC Site Report Chuck Boeheim Assistant Director SLAC Computer Services.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Sun FIRE Jani Raitavuo Niko Ronkainen. Sun FIRE 15K Most powerful and scalable Up to 106 processors, 576 GB memory and 250 TB online disk storage Fireplane.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Tripwire Enterprise Server – Getting Started Doreen Meyer and Vincent Fox UC Davis, Information and Education Technology June 6, 2006.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Backup Rationalisation Reorganisation of the CERN Computer Centre Backups David Asbury IT/DS Friday 6 December 2002.
CT NIKHEF June File server CT system support.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Richard P. Mount CHEP 2000Data Analysis for SLAC Physics Richard P. Mount CHEP 2000 Padova February 10, 2000.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
SLAC National Accelerator Laboratory Site Report A National Lab in Transition Randy Melen, Deputy CIO Computing Division, Operations Directorate SLAC National.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
CASPUR Site Report Andrei Maslennikov Sector Leader - Systems Catania, April 2001.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
May Richard P. Mount, SLAC Advanced Computing Technology Overview Richard P. Mount Director: Scientific Computing and Computing Services Stanford.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
CASPUR Site Report Andrei Maslennikov Lead - Systems Karlsruhe, May 2005.
W.A.Wojcik/CCIN2P3, May Running the multi-platform, multi-experiment cluster at CCIN2P3 Wojciech A. Wojcik IN2P3 Computing Center
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
23 April 2002HEP SYSMAN meeting1 Cambridge HEP Group - site report April 2002 John Hill.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
4-8 th October 1999CERN Site Report, HEPiX SLAC. A.Silverman CERN Site Report HEPNT/HEPiX October 1999 SLAC Alan Silverman CERN/IT/DIS.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
RAL Site report John Gordon ITD October 1999
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
14 th April 1999CERN Site Report, HEPiX RAL. A.Silverman CERN Site Report HEPiX April 1999 RAL Alan Silverman CERN/IT/DIS.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
May 10, 2000PHENIX CC-J Updates1 PHENIX CC-J Updates - Preparation For Opening - N.Hayashi / RIKEN May 10, 2000 PHENIX Computing
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Batch Software at JLAB Ian Bird Jefferson Lab CHEP February, 2000.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Oct. 6, 1999PHENIX Comp. Mtg.1 CC-J: Progress, Prospects and PBS Shin’ya Sawada (KEK) For CCJ-WG.
W.A.Wojcik/CCIN2P3, HEPiX at SLAC, Oct CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
PC Farms & Central Data Recording
Lee Lueking D0RACE January 17, 2002
Presentation transcript:

Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader

Randy MelenApril 14, Past 12 months... n Busy! n Target of May 9 for BaBar detector to begin n Challenge to get systems assembled and tested in time, to get C++ code working and sufficiently optimized, to handle 100 events/second for reconstruction and event recording n Once BaBar data begins, more difficult to make system changes, take service outages

Randy MelenApril 14, New Hardware Developments n Increased Solaris batch systems to compute farm ( from 5 Sun Ultra 2300 systems to 18 systems) n Upgraded Sun UE 6000 to 4GB memory n Acquired 4 Sun UE4500 systems, increased to 6 systems, for HPSS data movers, total of 4TB of disk n Acquired Sun UE10000 (24 CPUs, 12GB memory, 1.5TB disk, 2 domains) n 4 Sun E250 systems as tape movers n 3 IBM F50 systems as data movers

Randy MelenApril 14, New Hardware Developments (cont.) n Added 220 Sun U5 systems (256MB, 9GB IDE disk, 333MHz UltraSPARC IIi with 2MB cache, $188/SI95) n Expect to add ~200 more U5 systems 2Q1999, probably more disk, perhaps UE10000 upgrade to 400MHz CPUs

Randy MelenApril 14, Farm Management n Upgraded farm master for LSF to IBM F50 n Working with Sun Auto Client software and cacheFS to centrally manage Sun U5 systems n Actively doing Solaris performance tuning on UE6000 and UE10000 n Adding 2 Sun E250 systems as BaBar build systems; need to be able to build 1M C++ lines of code each night (twice?)

Randy MelenApril 14, Mass Storage Hardware n Upgraded 5 STK silos to PowderHorn robots n Added a 6’th STK silo and 12 STK Eagle drives; more Eagle drives will be needed n Need to add BaBar data import/export tape device; considering STK 9740 with DLT 7000 and RedWood drives

Randy MelenApril 14, Farm Network Technology n Currently using 3 Cisco Catalyst 5500 switches (~1.2 Gbps backplanes), everything on Fast Ethernet, single collision domains n Migrating to 3 Cisco Catalyst 6509 switches (~16 Gbps backplanes) n Deploying Gb Ethernet on ~16 Solaris servers

Randy MelenApril 14, HPSS Phase 3 (Porting) Ongoing n With assistance from Sun, began moving and testing the Solaris port to Solaris 2.6 n Lots of issues related to getting infrastructure pieces at correct version levels n Began HPSS 4.1 datamover port to Solaris 2.6 n Sun and IBM signed agreement for IBM to port HPSS 4.1A; we expect to deploy ~4Q1999

Randy MelenApril 14, HPSS Stage 4 (PRV0) Plans n While Solaris port continues, use IBM F50 systems as datamovers n Move development (porting and testing) to Solaris U250 build servers

Randy MelenApril 14, Currently Supported Systems n General Servers u generally Solaris > Solaris 2.6 u AFS servers will become Sun U2300 systems for AFS 3.5 multithreading u AIX > u phasing out “core” NFS file server (AIX 3.2.5!) by moving binaries and home directories to AFS n Farm Servers u AIX now frozen, not a porting platform for BaBar as of 7/1998 u Solaris > 2.6 completed n Desktop u still NT though much more Linux than before now

Randy MelenApril 14, Intel Farm Prototype n A prototype 17 node Intel compute farm acquired 4Q1998: u 2-way 256MB, 9GB disk, Dell 450MHz Pentium-II u partnership with Accelerator Research group and NERSC u strong interest in MPI and developing for Cray T3E production u decided on Linux from RedHat u modest success so far for scalability u expect to expand to 32 nodes 3Q1999 u Issues that remain: F Commercial software support (e.g., Objectivity, AFS, LSF with AFS support) F Manageability of large numbers of systems F MPI cluster vs “task farm”