Database Hardware Resources at Tier-1 Sites Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008.

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

IWR Ideen werden Realität Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Institut für Wissenschaftliches Rechnen Status of Database Services.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site Adam Padee ( ) Ryszard Gokieli ( ) Krzysztof.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Oracle Clustering and Replication Technologies CCR Workshop - Otranto Barbara Martelli Gianluca Peco.
INFN-T1 site report Giuseppe Misurelli On behalf of INFN-T1 staff HEPiX Spring 2015.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
1 RAL Status and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, CERN, November 2009.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
Castor F2F Meeting Barbara Martelli Castor Database CNAF.
Storage review Carmine Cioffi Database Administrator and Developer 3D Workshop, Barcelona (ES), April 2009.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Experience and Lessons learnt from running High Availability Databases on Network Attached Storage Ruben Gaspar Manuel Guijarro et al IT/DES.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Corporate Partner Overview and Update September 27, 2007 Gary Crane SURA Director IT Initiatives.
CERN IT Department CH-1211 Geneva 23 Switzerland t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben.
ASGC 1 ASGC Site Status 3D CERN. ASGC 2 Outlines Current activity Hardware and software specifications Configuration issues and experience.
Reliable Datagram Sockets and InfiniBand Hanan Hit NoCOUG Staff 2010.
Service Review and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, Barcelona (ES), April 2009.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
WLCG Service Report ~~~ WLCG Management Board, 27 th October
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
WLCG Service Report ~~~ WLCG Management Board, 24 th November
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
CASTOR Databases at RAL Carmine Cioffi Database Administrator and Developer Castor Face to Face, RAL February 2009.
NL Service Challenge Plans Kors Bos, Sander Klous, Davide Salomoni (NIKHEF) Pieter de Boer, Mark van de Sanden, Huub Stoffers, Ron Trompert, Jules Wolfrat.
LFC Replication Tests LCG 3D Workshop Barbara Martelli.
Summary of the status and plans of FTS and LFC back-end database installations at Tier-1 Sites Gordon D. Brown Rutherford Appleton Laboratory WLCG Workshop.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
Review of Recent CASTOR Database Problems at RAL Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
News from Alberto et al. Fibers document separated from the rest of the computing resources
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
CASTOR CNAF TIER1 SITE REPORT Geneve CERN June 2005 Ricci Pier Paolo
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
PIC port d’informació científica DateText1 November 2009 (Elena Planas) PIC Site review.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
PIC port d’informació científica Luis Diaz (PIC) ‏ Databases services at PIC: review and plans.
Database Project Milestones (+ few status slides) Dirk Duellmann, CERN IT-PSS (
LCG Tier1 Reliability John Gordon, STFC-RAL CCRC09 November 13 th, 2008.
HP Proliant Server  Intel Xeon E3-1220v3 (3.1GHz / 4-core / 8MB / 80W).  HP 4GB Dual Rank x8 PC E (DDR3-1600) Unbuffered Memory Kit.  HP Ethernet.
TRIUMF Site Report for HEPiX, JLAB, October 9-13, 2006 – Corrie Kost TRIUMF SITE REPORT Corrie Kost & Steve McDonald Update since Hepix Spring 2006.
RAL Plans for SC2 Andrew Sansum Service Challenge Meeting 24 February 2005.
1062m0656 between 10692m2192 DS/ICI/CIF EqualLogic PS6510E
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
Oracle Clustering and Replication Technologies UK Metadata Workshop - Oxford Barbara Martelli Gianluca Peco.
A Nordic Tier-1 for LHC Mattias Wadenstein Systems Integrator, NDGF Grid Operations Workshop Stockholm, June the 14 th, 2007.
A Distributed Tier-1 for WLCG Michael Grønager, PhD Technical Coordinator, NDGF CHEP 2007 Victoria, September the 3 rd, 2007.
NDGF Site Report Mattias Wadenstein Hepix 2009 spring, Umeå , Umeå University.
KOLKATA Grid Kolkata Tier-2 Status and Plan Site Name :- IN-DAE-VECC-02 Gocdb Name:- IN-DAE-VECC-02 VO :- ALICE City:- KOLKATA Country :-
ICEPP, University of Tokyo
Dirk Duellmann CERN IT/PSS and 3D
Database Services at CERN Status Update
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Oracle Storage Performance Studies
ASM-based storage to scale out the Database Services for Physics
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Presentation transcript:

Database Hardware Resources at Tier-1 Sites Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008

Overview Database Machines Storage Plans Discussion

Nodes in Production Cluster 3D Atlas3D LHCbFTSLFCCASTORSRM CERN3432 CA-TRIUMF21 DE-GridKa2 ES-PIC32 FR-IN2P3 IT-CNAF32L2L 3A3A 2 L /3 A 4 NDGF1 NL-SARA2 TW-ASGC3232 UK-RAL32310 US-BNL4

Hardware Manufacturer CERNDell PowerEdge 2950 CA-TRIUMFHP ProLiant DL380 G5, Dell PowerEdge 1950 DE-GridKaIBM x336 ES-PICFujitsu PRIMERGY BX600 S3 FR-IN2P3 IT-CNAFDell PowerEdge 1950/2950 NDGFDell PowerEdge 1950 NL-SARADell 1850 TW-ASGCQuanta Blade Server, QB600 Xeon_E7520Server Blade UK-RALSupermicro US-BNL

DB Machines ProcCoreHyperSpeedMemDiskRedundancy CERN24No2.33GHz16GB476GB x 2dual power, mirrored disks, 4 NIC (2 private/2 public), dual HBA CA-TRIUMF22No3.00GHz10GB73GBDual power, RAID, switches DE-GridKa22No3.2GHz4GB73GBNo ES-PIC22?1.6GHz8GB75GBDual power, three ethernet FR-IN2P3 IT-CNAF22Yes3.2GHz4GB100GBDual power NDGF22No3.00GHz4GB72GBDual power NL-SARA22Yes3.2 GHz4GB140GBDual Power TW-ASGC21No3.0 GHz8GB80 GBN+1 power supplies UK-RAL22No2.4GHz4GB250GBDual power US-BNL223.0GHZ16GB

Storage TypeModelRAIDRawAfterRedundancy CERNSAN1+0 (ASM) 44.8TB~20TBDual channel CA-TRIUMFSANHP MSA20ASM4.5TB2.1TBFC switches, n/w switches, dual-port FC cards DE-GridKaSANCondor62TBNo ES-PICSANNetApp – FAS3040Double parity 18TB~6TBDual channel, dual controllers FR-IN2P3 IT-CNAFSANEMC CX3-801/54TB2.3TBFull, dual-port FC cards NDGFSANEMC CX7001/53.2TB2TB2 networks and controllers NL-SARASANSGI TP 91005n/a700GBDual channel TW-ASGCSANInfortrend A24F-G TB15TBSingle controller UK-RALSANInfortrend A16F-G TB1.75TB US-BNLSANIBM DS34006TB

CERN CASTOR Hardware Setup –2 nodes CASTOR name server –6 nodes CASTOR+SRM ATLAS (2 stager, 2 DLF, 2 SRM) –6 nodes CASTOR+SRM ALICE (2 stager, 2 DLF, 2 SRM) –6 nodes CASTOR+SRM CMS (2 stager, 2 DLF, 2 SRM) –6 nodes CASTOR+SRM LHCb (2 stager, 2 DLF, 2 SRM) –6 nodes CASTOR+SRM "Public" (2 stager, 2 DLF, 2 SRM) –a few others for "ITDC" environment (testing) / dev / test Hardware manufacturer –HP and Dell

CERN CASTOR Hardware Machine model –HP DL380 G5 and Dell 1950 Number of processors per machine –2 Number of cores per processer –2 for the HPs and 4 for the Dells Processor clock speed –2.33GHz Hyperthreading –No

CERN CASTOR Hardware Memory size –8 or 16GB Local disk size –73GB Redundancy –dual power supplies –mirrored local disks –5 NIC (2 storage, 2 private, 1 public) Clustered storage type –NAS

CERN CASTOR Hardware Storage manufacturer –NetApp 3020 and 3040 Storage RAID (0, 1, 5, 1+0 etc) –RAID-DP (double parity) Storage space (raw) –Fibre Channel disks (10k rpm) and SATA disks –41.1TB Storage space (after RAID) –32.3TB Redundancy (dual channel etc) –Two paths to storage

Hardware Plans GridKa –New server in Spring 2009 PIC –None Triumf –3D,FTS - add a new 2-node cluster and storage by April 2009 –LFC - monitor performance, review situation at least each year, before renewing hardware support contract. NDGF –We are planning and have been planning to have a 3-node cluster for some time now. It looks like the storage system will be SAN with 4Gbit/s FC and SAS disks. Nodes will be something adequate, maybe dual Quad-core Xeons with 8 or 16GB of RAM.

Hardware Plans CNAF –Just installed: 10TB of SATA storage (for flash recovery area) and10 TB of FibreChannel storage (4Gbit) The Storage device is the same CX3-80 mentioned before. This new storage will be mainly used for INFN services, but some of it can be allocated to 3D or LHC experiments clusters if needed. CERN (CASTOR) –renewal of the HPs, not decided yet, blades likely SARA –We're in the process of purchasing new hardware.This process should be completed within a few months.

Hardware Plans ASGC –as outline in server spec, we plan to add new instances into same rac group but base on different profile from which the remote management is more stable and robust enough comparing with current solution adopt. CERN (3D) –probably to DELL blade systems, storage not known yet. RAL –SAN redundancy –failover on CASTOR

Group Questions When do you replace hardware? Is the 3D kit old now? Priorities for redundancy? Switch redundancy? How much is DBA involved in procurement? How much does sysadmin know about databases? Is warranty important?

Questions and (hopefully) Answers