ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, 2000-3000 in this year  Distributed.

Slides:



Advertisements
Similar presentations
4/11/2017 © 2014 Microsoft Corporation. All rights reserved. Microsoft, Windows, and other product names are or may be registered trademarks and/or trademarks.
Advertisements

♦ Commodity processor with commodity inter- processor connection Clusters Pentium, Itanium, Opteron, Alpha GigE, Infiniband, Myrinet, Quadrics, SCI NEC.
2. Computer Clusters for Scalable Parallel Computing
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
Program Systems Institute Russian Academy of Sciences1 Program Systems Institute Research Activities Overview Extended Version Alexander Moskovsky, Program.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Information Technology Center Introduction to High Performance Computing at KFUPM.
AASPI Software Computational Environment Tim Kwiatkowski Welcome Consortium Members November 18, 2008.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
HIGH PERFORMANCE COMPUTING ENVIRONMENT The High Performance Computing environment consists of high-end systems used for executing complex number crunching.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
HELICS Petteri Johansson & Ilkka Uuhiniemi. HELICS COW –AMD Athlon MP 1.4Ghz –512 (2 in same computing node) –35 at top500.org –Linpack Benchmark 825.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
An Introduction to Princeton’s New Computing Resources: IBM Blue Gene, SGI Altix, and Dell Beowulf Cluster PICASso Mini-Course October 18, 2006 Curt Hillegas.
CSC Site Update HP Nordic TIG April 2008 Janne Ignatius Marko Myllynen Dan Still.
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
Real Parallel Computers. Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra, Meuer, Simon Parallel.
1 petaFLOPS+ in 10 racks TB2–TL system announcement Rev 1A.
Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Building a High-performance Computing Cluster Using FreeBSD BSDCon '03 September 10, 2003 Brooks Davis, Michael AuYeung, Gary Green, Craig Lee The Aerospace.
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
Reliable Datagram Sockets and InfiniBand Hanan Hit NoCOUG Staff 2010.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Diamond Computing Status Update Nick Rees et al..
Experience with the Thumper Wei Yang Stanford Linear Accelerator Center May 27-28, 2008 US ATLAS Tier 2/3 workshop University of Michigan, Ann Arbor.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
Infiniband in EDA (Chip Design) Glenn Newell Sr. Staff IT Architect Synopsys.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Computing Resources at Vilnius Gediminas Technical University Dalius Mažeika Parallel Computing Laboratory Vilnius Gediminas Technical University
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
News from Alberto et al. Fibers document separated from the rest of the computing resources
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
PIC port d’informació científica DateText1 November 2009 (Elena Planas) PIC Site review.
IHEP Computing Site Report Shi, Jingyan Computing Center, IHEP.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Operational and Application Experiences with the Infiniband Environment Sharon Brunett Caltech May 1, 2007.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
HELMHOLTZ INSTITUT MAINZ Dalibor Djukanovic Helmholtz-Institut Mainz PANDA Collaboration Meeting GSI, Darmstadt.
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The Naples' testbed for the SuperB computing model: first tests G. Russo, D. Del Prete, S.
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
E VALUATION OF GLUSTER AT IHEP CHENG Yaodong CC/IHEP
© 2007 Z RESEARCH Z RESEARCH Inc. Non-stop Storage GlusterFS Cluster File System.
Brief introduction about “Grid at LNS”
NIIF HPC services for research and education
HPC Roadshow Overview of HPC systems and software available within the LinkSCEEM project.
Experience of Lustre at QMUL
Enterprise Storage at Seton Hall University
The demonstration of Lustre in EAST data system
Experience of Lustre at a Tier-2 site
Stallo: First impressions
Low-Cost High-Performance Computing Via Consumer GPUs
Procurements at CERN: Status and Plans
System G And CHECS Cal Ribbens
Overview of HPC systems and software available within
Cluster Computers.
Presentation transcript:

ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed parallel filesystem of 1 PB in 3 years, TB in this year  kW of power (35-40 kW in this year)

ITEP computing center and plans for supercomputing

Hardware:  7U blade system with 10 twin modules (20 nodes per blade chassis)  3 blade chassis enclosures with power supply per 42U rack kW per 42U rack.  Two 36 QSFP ports infiniband switches per blade chassis  36 QSFP porrts infiniband switches for the second level of fat tree  2 x AMD 12 cores CPU per node  64 GB of RAM per node  Two channel of 4xQDR Infiniband per node for interprocess communication ITEP computing center and plans for supercomputing

Infiniband topology: Two levels fat tree using 36 x QSFP ports infiniband switches

ITEP computing center and plans for supercomputing Software: RedHat based distribution (Scientific Linux or CentOS) for x86_64 architecture. TORQUE batch system with maui as scheduler. OpenMPI with TORQUE integration (mvapich and mvapich2 is under consideration) OpenMP BLAS, lapack including ATLAS versions, ACML

ITEP computing center and plans for supercomputing Prototype:  7U blade system  10 twin blade modules, 20 nodes  36 QSFP ports switch module  22x1GbE + 3x10GbE ports Ethernet switch module  Node characteristics:  Dual Xeon X GHz 6 cores  32GB RAM  500 GB disk  One 4xQDR Infiniband port  Dual 1GbE Ethernet (one channel connected)

ITEP computing center and plans for supercomputing Prototype software configuration:  CentOS 5.6 x86_64  TORQUE batch system  maui scheduler  OpenMPI 1.4 integrated with TORQUE  BLAS, lapack including ATLAS version

ITEP computing center and plans for supercomputing Benchmarking:  Single node (12 processes):  Linpack (N=60000, NB=128, P=4, Q=3)- 98 Gflops (77% of theoretical performance)  One process per node (12 processes):  Linpack (N=60000, NB=128, P=4, Q=3)- 100 Gflops (78% of theoretical performance)  Cluster full load (240 processes):  Linpack (N=250000, NB=128, P=16, Q=15) Gflops (70% of theoretical performance)

ITEP computing center and plans for supercomputing Distributed parallel filesystem:  Glusterfs with RDMA and TCP as transport  Local disks of the nodes are used for glusterfs volumes  Replication (mirroring) of data provides fault tolerance  Linear speed for 1 file:Write: 46 MB/s, Read: 84 MB/s  Bandwidth for cluster: Write: 450 MB/s, Read: 839 MB/s

ITEP computing center and plans for supercomputing Infiniband bandwidth measurements for prototype: Approximated bandwidth for cluster: 56 GByte/s for 20 nodes (bidirectional)

ITEP computing center and plans for supercomputing Zabbix monitoring system:  Active client with vast possibility of customization  SNMP monitoring and traps  IPMI monitoring and control  Triggers and events for group of hosts including the usage of the aggregate functions  Powerful and flexible tools for triggers and actions description  Presentation of data in many ways

ITEP computing center and plans for supercomputing Example of presentation data from different sources in ZABBIX

ITEP computing center and plans for supercomputing Next steps:  Dedicated storage with lustre filesystem  First stage of the supercomputer with AMD processors  Two levels fat tree infiniband topology  Moving to RHEL 6 based operating system

ITEP computing center and plans for supercomputing Thank you