PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Monte-Carlo method and Parallel computing  An introduction to GPU programming Mr. Fang-An Kuo, Dr. Matthew R. Smith NCHC Applied Scientific Computing.
DOSAR Workshop VI April 17, 2008 Louisiana Tech Site Report Michael Bryant Louisiana Tech University.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
Early Linpack Performance Benchmarking on IPE Mole-8.5 Fermi GPU Cluster Xianyi Zhang 1),2) and Yunquan Zhang 1),3) 1) Laboratory of Parallel Software.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Supermicro © 2009Confidential HPC Case Study & References.
Jia Yao Director: Vishwani D. Agrawal High Performance Compute Cluster April 13,
Tamnun Hardware. Tamnun Cluster inventory – system Login node (Intel 2 E GB ) – user login – PBS – compilations, – YP master Admin.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Computing Resources Joachim Wagner Overview CNGL Cluster MT Group Cluster School Cluster Desktop PCs.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
1 petaFLOPS+ in 10 racks TB2–TL system announcement Rev 1A.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Corporate Partner Overview and Update September 27, 2007 Gary Crane SURA Director IT Initiatives.
Florida Tier2 Site Report USCMS Tier2 Workshop Fermilab, Batavia, IL March 8, 2010 Presented by Yu Fu For the Florida CMS Tier2 Team: Paul Avery, Dimitri.
Site Lightning Report: MWT2 Mark Neubauer University of Illinois at Urbana-Champaign US ATLAS Facilities UC Santa Cruz Nov 14, 2012.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
High Performance Computing Processors Felix Noble Mirayma V. Rodriguez Agnes Velez Electric and Computer Engineer Department August 25, 2004.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Sandor Acs 05/07/
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
Weekly Report By: Devin Trejo Week of June 21, 2015-> June 28, 2015.
Short introduction Pulsar Parkes. Outline PDFB – Single beam pulsar timing system CASPER – Single beam pulsar coherent dedispersion system.
Computational Research in the Battelle Center for Mathmatical medicine.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
US ATLAS Western Tier 2 Status Report Wei Yang Nov. 30, 2007 US ATLAS Tier 2 and Tier 3 workshop at SLAC.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
ATLAS Midwest Tier2 University of Chicago Indiana University Rob Gardner Computation and Enrico Fermi Institutes University of Chicago WLCG Collaboration.
Florida Tier2 Site Report USCMS Tier2 Workshop Livingston, LA March 3, 2009 Presented by Yu Fu for the University of Florida Tier2 Team (Paul Avery, Bourilkov.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
PDSF and the Alvarez Clusters Presented by Shane Canon, NERSC/PDSF
What’s Coming? What are we Planning?. › Better docs › Goldilocks – This slot size is just right › Storage › New.
Ole’ Miss DOSAR Grid Michael D. Joy Institutional Analysis Center.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
Parallel Computers Today Oak Ridge / Cray Jaguar > 1.75 PFLOPS Two Nvidia 8800 GPUs > 1 TFLOPS Intel 80- core chip > 1 TFLOPS  TFLOPS = floating.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
S. Pardi Computing R&D Workshop Ferrara 2011 – 4 – 7 July SuperB R&D on going on storage and data access R&D Storage Silvio Pardi
Sobolev(+Node 6, 7) Showcase +K20m GPU Accelerator.
BeStMan/DFS support in VDT OSG Site Administrators workshop Indianapolis August Tanya Levshina Fermilab.
KOLKATA Grid Kolkata Tier-2 Status and Plan Site Name :- IN-DAE-VECC-02 Gocdb Name:- IN-DAE-VECC-02 VO :- ALICE City:- KOLKATA Country :-
Brief introduction about “Grid at LNS”
Experience of Lustre at QMUL
GRID OPERATIONS IN ROMANIA
Low-Cost High-Performance Computing Via Consumer GPUs
Belle II Physics Analysis Center at TIFR
HEPSYSMAN Summer June 2017 Ian Loader Chris Brew
Update on Plan for KISTI-GSDC
Experience of Lustre at a Tier-2 site
Southwest Tier 2.
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Tamnun Hardware.
CUBAN ICT NETWORK UNIVERSITY COOPERATION (VLIRED
HEPSYSMAN Summer June 2017 Ian Loader Chris Brew
Presentation transcript:

PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy Philpott, JLAB)

Outline PDSF Structure Compute Node H/W Storage H/W Networking and Data Transfer PDSF Grid Infrastructure Other NERSC News

PDSF Structure PDSF structure has not changed. We still have a total of 2.5 FTEs working on PDSF All nodes are now 64-bit systems (AMD and Intel) – 1 st gen Opteron, Barcelona – Xeon 5400, Nehalem New purchases – Compute nodes, quad-core Nehalems – Storage FC/SATA, Dell MD3K – Networking – 10GE switch Dell 8024

Compute Node H/W New compute nodes – Quad-core, dual-socket, Intel ONLY – Nodes have 3GB RAM per core – HEP-SPEC06 of ~92 (8core), ~115 (16core) – For now, running 8 tasks per node, investigating ways of running more (memory limitations of jobs prevent using 16 slots) Started using 2U-Twin nodes (Supermicro) for admin nodes. Evaluating Dell 6100C (also a 2U-Twin)

Storage H/W Continuing with a mix of F/C-SATA storage and SAS-SATA storage. Evaluating some “integrated” storage, i.e. server, RAID controller and disks all in one chassis for non-GPFS XRootD servers.

Networking and Data Transfer Evaluating the new 10GE switch from Dell (PC 8024F) – Uses SFP+ connectors Continue with Dell 6200 switches for top-of-the-rack (GigE to hosts and 10GE uplink) Providing a “Data Transfer Node” service to users. – High bandwidth connection to the outside world – 10 Gigabit connection to NERSC backbone and internal NERSC – High bandwidth connection to local PDSF storage – Network stack tuned for large-scale transfers

PDSF Grid Infrastructure One gatekeeper OSG – Reporting as NERSC-PDSF Two storage elements OSG Bestman – SRM servers NERSC-PDSF and NERSC-DTN – Help with testing of Bestman for the LBNL dev. team – Bestman buffering on a disk local to the server (2TB cache) data transfers from the cluster to archival storage. This allows jobs to finish without waiting on the batch nodes for the transfers to complete. Prototype Alice VO box feasibility test conducted in December (Alien submits Condor-G jobs to OSG gatekeepers) – Passed scalability tests – Alice vo box being deployed - testing under way

Dirac GPU Cluster 48 nodes –Dual 2.4GHz Nehalem 5500 processors –24GB memory –1 TB drive –Dual PCIe x16 slots ConnectX2 QDR Tesla GPU soon to be Tesla ‘Fermi’ GPU Resource available from the Carver (IBM iDataPlex) cluster

Tesla GPUs C CUDA cores 4 GB memory 800 MHz memory speed 78 Gflops peak double precision floating point C CUDA cores ECC Memory 3 GB memory –6 GB for C GHz memory speed 515 Gflops peak double precision floating point

IPv6 testing at NERSC NERSC has started testing the use of IPv6 We have brought up an IPv6 peering with ESnet and are testing and modifying our security systems. – We are still looking for a scientific customer use of IPv6 – We will create a roadmap to deliver IPv6 according to need – First likely requirement will be a cluster that needs more than 8k public addresses.