CASTOR Databases at RAL Carmine Cioffi Database Administrator and Developer Castor Face to Face, RAL 18-19 February 2009.

Slides:



Advertisements
Similar presentations
2 Copyright © 2005, Oracle. All rights reserved. Installing the Oracle Database Software.
Advertisements

DB server limits (process/sessions) Carlos Fernando Gamboa, BNL Andrew Wong, TRIUMF WLCG Collaboration Workshop, CERN Geneva, April 2008.
Wim Coekaerts Director of Linux Engineering Oracle Corporation.
Oracle Clustering and Replication Technologies CCR Workshop - Otranto Barbara Martelli Gianluca Peco.
CASTOR Upgrade, Testing and Issues Shaun de Witt GRIDPP August 2010.
Database Hardware Resources at Tier-1 Sites Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Introduction to Oracle Backup and Recovery
1 RAL Status and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, CERN, November 2009.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Castor F2F Meeting Barbara Martelli Castor Database CNAF.
Storage review Carmine Cioffi Database Administrator and Developer 3D Workshop, Barcelona (ES), April 2009.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Fermilab Oct 17, 2005Database Services at LCG Tier sites - FNAL1 FNAL Site Update By Anil Kumar & Julie Trumbo CD/CSS/DSG FNAL LCG Database.
ASGC 1 ASGC Site Status 3D CERN. ASGC 2 Outlines Current activity Hardware and software specifications Configuration issues and experience.
Luca Canali, CERN Dawid Wojcik, CERN
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Backup & Recovery Backup and Recovery Strategies on Windows Server 2003.
Service Review and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, Barcelona (ES), April 2009.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
WLCG Service Report ~~~ WLCG Management Board, 27 th October
Copyright © 2013 Scale Abilities Ltd … to Oracle 11.2 on Linux From Oracle 11.1 on Solaris …
Operation of CASTOR at RAL Tier1 Review November 2007 Bonny Strong.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
A Guide to Oracle9i1 Database Instance startup and shutdown.
DB Installation and Care Carmine Cioffi Database Administrator and Developer ICAT Developer Workshop, The Cosener's House August
CERN Physics Database Services and Plans Maria Girone, CERN-IT
Continuous DB integration testing with RAT „RATCOIN”
DB Questions and Answers open session Carlos Fernando Gamboa, BNL WLCG Collaboration Workshop, CERN Geneva, April 2008.
Report from CASTOR external operations F2F meeting held at RAL in February Barbara Martelli INFN - CNAF.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
Review of Recent CASTOR Database Problems at RAL Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008.
Queensland University of Technology CRICOS No J VMware as implemented by the ITS department, QUT Scott Brewster 7 December 2006.
CERN IT Department CH-1211 Genève 23 Switzerland t Load Testing Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th 2009.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Oracle 10g Automatic Storage Management Overview of ASM as a Storage Option for Oracle 10g.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Database Competence Centre openlab Major Review Meeting nd February 2012 Maaike Limper Zbigniew Baranowski Luigi Gallerani Mariusz Piorkowski Anton.
LHCb File-Metadata: Bookkeeping Carmine Cioffi Department of Physics, Oxford University UK Metadata Workshop Oxford, 04 July 2006.
Distributed Logging Facility Castor External Operation Workshop, CERN, November 14th 2006 Dennis Waldron CERN / IT.
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
PIC port d’informació científica DateText1 November 2009 (Elena Planas) PIC Site review.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
ASGC Site Report Jason Shih ASGC Grid Ops CASTOR External Operation Face to Face Meeting.
CASTOR Status at RAL CASTOR External Operations Face To Face Meeting Bonny Strong 10 June 2008.
LHC Logging Cluster Nilo Segura IT/DB. Agenda ● Hardware Components ● Software Components ● Transparent Application Failover ● Service definition.
FroNtier Stress Tests at Tier-0 Status report Luis Ramos LCG3D Workshop – September 13, 2006.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
PIC port d’informació científica Luis Diaz (PIC) ‏ Databases services at PIC: review and plans.
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Bonny Strong RAL RAL CASTOR Update External Institutes Meeting Nov 2006 Bonny Strong, Tim Folkes, and Chris Kruk.
Considerations for database servers Castor review – June 2006 Eric Grancher, Nilo Segura Chinchilla IT-DES.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
3 Copyright © 2006, Oracle. All rights reserved. Installation and Administration Basics.
Oracle 10g database installation kit  A bundle of scripts which allows to install Oracle 10g database server on a single node: Useful for both experienced.
DB Questions and Answers open session (comments during session) WLCG Collaboration Workshop, CERN Geneva, 24 of April 2008.
14 Copyright © 2007, Oracle. All rights reserved. Backup and Recovery Concepts.
Oracle 10g Administration Database Architecture, Creation and Interfaces Copyright ©2006, Custom Training Institute.
Oracle Clustering and Replication Technologies UK Metadata Workshop - Oxford Barbara Martelli Gianluca Peco.
Database Services at CERN Status Update
NGS Oracle Service.
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Ákos Frohner EGEE'08 September 2008
Oracle Database Monitoring and beyond
Oracle Storage Performance Studies
ASM-based storage to scale out the Database Services for Physics
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Presentation transcript:

CASTOR Databases at RAL Carmine Cioffi Database Administrator and Developer Castor Face to Face, RAL February 2009

Database Services Team Gordon Brown (not the prime minister!!!) Keir Hawker Richard Sinclair Carmine Cioffi 2

Outline Database Overview Schemas Size and Versions Oracle Installation Hardware Overview Hardware Specification Hardware Limitation and Future Plan New Hardware Architecture How we will move over the new hardware Problems hit during production Test Database Tuning and Monitoring Few Metrics Knowledge Exchange 3

Databases Overview Two 5 nodes RAC + one single instance in production: Pluto Neptune Uranus One two nodes RAC for development and testing: Vulcan 4

Database Overview 5 NeptuneVulcanPluto Switch 1 Disk array Switch 2 Uranus SCSI attached disk array

Databases Overview Our databases are named after Roman Gods: Pluto: god of the underworld and the judge of the dead. Neptune: God of sea Uranus: god of the sky/heaven Vulcan: god of fire and manufacturer of arms 6

Databases Overview Each database host the following schemas: Pluto: Name Server, VMGR (Volume Manager), CUPV (Castor User Privilege Validator), CMS Stager, Gen Stager, Repack, SRM Alice, SRM CMS Neptune: Atlas Stager, LHCb stager, SRM Atlas, SRM LHCb. Uranus: DLF for all VOs 7

Schemas Size and Versions 8 SchemasVersionSize Name Servern/a2GB VMGRn/a1.5MB CUPVn/a0.2MB CMS Stager2_1_7_19_21GB Gen Stager2_1_7_19_22.5GB Repack2_1_7_1924MB SRM Alice1_1_0556M B SRM CMS2_7_12155M B Pluto SchemasVersionSize Atlas Stager2_1_7_19_2134GB LHCb stager2_1_7_19_2600M B SRM Atlas2_7_123GB SRM LHCb2_7_1276MB Neptune

Oracle Installation Version Last patch CPUJan09 Non default initialisation parameters: –_kks_use_mutex_pin = FALSE –sga_target = 2GB –pga_aggregate_target = 600M –processes = 800 –cursor_sharing = EXACT –open_cursors = 300 9

Hardware Overview 10 NeptuneVulcanPluto FC switch 1 Brocade 200E 4Gbit Disk array Overland 1200 Pluto Vulcan Neptune OCR and VD for the three RACs FC switch 2 Brocade 200E 4Gbit Uranus SCSI attached disk array

Hardware Specification OS: Red Hat Enterprise Linux AS release 4 (Nahant Update 7) RAM: 4GB CPU: Dual quad Intel(R) Xeon(TM) 3.00GHz Storage: –Pluto: 560GB –Neptune: 560GB –Vulcan:93GB –Uranus: 1.8TB Overland 1200 disk array –twin controller –twin Fibre Channel ports to each controller –10 SAS disks (300GB each 3TB total gross space) –Raid 1(1.5 TB net space) Two Brocade 200E 4Gbit switched 11

Hardware Limitation and Future Plan A concern is raised by the current hardware : –There is not enough storage –There is no disk array redundancy: if the disk array breaks down the Castor database goes down Because of the above concern we are planning to move over to new hardware: –Two new disk arrays with more space –One disk array will be used to mirror the other one using Oracle ASM 12

New Hardware Architecture 13 Neptune PlutoVulcan FC switch 1 Brocade 200E 4Gbit Disk array 2 EMC Clairion AX4-5 Pluto Neptune OCR and VD for the two RACs FC switch 2 Brocade 200E 4Gbit Disk array 1 EMC Clairion AX4-5 Pluto Neptune FC switch Brocade 200E 4Gbit Disk array 2 EMC Clairion AX4-5 Vulcan Uranus SCSI attached disk array OCR and VD ASM Mir.

New Hardware Specifications Storage: –Pluto: 2TB –Neptune: 3TB Two Overland 1200 disk array each one with –twin controller –twin Fibre Channel ports to each controller –20 SAS disks (300GB each 6TB total gross space) Two Brocade 200E 4Gbit switched Raid 5 (5TB total net space) 14

How we will move over the new hardware The current ASM disk group has redundancy external which cannot be modified therefore we will proceed in this way: –Create a new disk group with redundancy normal –Shutdown the database and migrate data from the current disk group to the new disk group. –Adjust the control file content so that the tablespaces are associate with the data files in the new location –Start-up the database 15

Problems hit during production 1.Big Id when updating the Id2Type table: –It does happen randomly and we are trying to recreate it on demand so that Oracle support can investigate it 2.Cross talk experienced during the synchronization process: –was resolved by removing the synchronisation process from Castor. We are trying to recreate it over the testing system 3.Atlas stager tables stats becomes stale because of high throughput: –When the problem appeared we recollected the statistics. We are changing the script to collect statistics similar to CERN (i.e. separate statistics gathering scripts for internal Oracle tables and Castor tables. 16

Problems hit during production 1.Running out of space because of the huge amount of redo logs generated (~300GB day) 2.RMAN is very slow to work over NFS mounting points: –The backups were done over an NFS mounting point and this caused RMAN to be very slow in removing redo logs. Because Oracle could create redo logs faster than RMAN could remove them many times we got close to fill up the ASM. We solved the problem by doing backup over locally attached disks which sped up RMAN 17

Test Database Vulcan is used for testing Oracle version Oracle Software not updated/patched It is used to try to recreate the: –Big Id problem –Cross talk problem Schema version and size: 18 SchemasVersionSize SRM ATLAS1_1_096MB Atlas Stager2_1_7_1516MB LHCb stager2_1_7_1516MB

Tuning and Monitoring Grid Control AWR CPU load (top) Oracle advisors Explain execution plan 19

Few Metrics 350 tran/sec on Neptune 290 tran/sec for Atlas Stager 780 physical write/ sec (6MB/sec) on Neptune 700 physical write/ sec (5.5MB/sec) for Atlas Stager 930 physical read/sec ( 7.3MB/sec) on Neptune 910 physical read/sec ( 7.15MB/sec) for Atlas Stager Orion results: –MBPS= –IOPS=1652 –Latency msec=

Knowledge Exchange Do we want to have regular Castor database workshops? Exchange on regular basis Metrics? –I/O sec –Tran/sec –AWR reports I would like to see the system architectures for Tier(0/1)s sites I found useful the Castor time slot into the 3D fortnight meeting (even though I’m the only one to report problems ;o) Any suggestion? 21