EGO Computing Center site report EGO - Via E. Amaldi 56021 S. Stefano a Macerata - Cascina (PI) | Stefano Cortese INFN Computing Workshop – 26-05-2004.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
A new standard in Enterprise File Backup. Contents 1.Comparison with current backup methods 2.Introducing Snapshot EFB 3.Snapshot EFB features 4.Organization.
Beowulf Supercomputer System Lee, Jung won CS843.
LANs and WANs Network size, vary from –simple office system (few PCs) to –complex global system(thousands PCs) Distinguish by the distances that the network.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
UC Berkeley 1 A Disk and Thermal Emulation Model for RAMP Zhangxi Tan and David Patterson.
Teraserver Darrel Sharpe Matt Todd Rob Neff Mentor: Dr. Palaniappan.
VMware Infrastructure Alex Dementsov Tao Yang Clarkson University Feb 28, 2007.
Deployment Options Frank Bergmann
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
D-Link International Private Limited Training and Staff Development Department Module : Network Attached Storage Module : Network Attached Storage.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
Types of Operating System
Module 10 Configuring and Managing Storage Technologies.
Day 10 Hardware Fault Tolerance RAID. High availability All servers should be on UPSs –2 Types Smart UPS –Serial cable connects from UPS to computer.
Operational computing environment at EARS Jure Jerman Meteorological Office Environmental Agency of Slovenia (EARS)
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
1 © 2010 Overland Storage, Inc. © 2012 Overland Storage, Inc. Overland Storage The Storage Conundrum Neil Cogger Pre-Sales Manager.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Course ILT Basics of information technology Unit objectives Define “information technology” (IT), distinguish between hardware and software, and identify.
Confidential1 Introducing the Next Generation of Enterprise Protection Storage Enterprise Scalability Enhancements.
Server Systems Administration. Types of Servers Small Servers –Usually are PCs –Need a PC Server Operating System (SOS) such as Microsoft Windows Server,
Paul Scherrer Institut 5232 Villigen PSI HEPIX_AMST / / BJ95 PAUL SCHERRER INSTITUT THE PAUL SCHERRER INSTITUTE Swiss Light Source (SLS) Particle accelerator.
MURI Hardware Resources Ray Garcia Erik Olson Space Science and Engineering Center at the University of WI - Madison.
Guide to Linux Installation and Administration, 2e1 Chapter 2 Planning Your System.
CASPUR Site Report Andrei Maslennikov Lead - Systems Karlsruhe, May 2005.
CS/IS 465: Data Communication and Networks 1 CS/IS 465 Data Communications and Networks Lecture 28 Martin van Bommel.
1 Selecting LAN server (Week 3, Monday 9/8/2003) © Abdou Illia, Fall 2003.
Block1 Wrapping Your Nugget Around Distributed Processing.
TRIUMF Site Report for HEPiX/HEPNT, Vancouver, Oct20-24/2003 – Corrie Kost TRIUMF SITE REPORT Corrie Kost Head Scientific Computing.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
Spending Plans and Schedule Jae Yu July 26, 2002.
ITGS Networks. ITGS Networks and components –Server computers normally have a higher specification than regular desktop computers because they must deal.
The DCS lab. Computer infrastructure Peter Chochula.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
Monte Carlo Data Production and Analysis at Bologna LHCb Bologna.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Queensland University of Technology CRICOS No J VMware as implemented by the ITS department, QUT Scott Brewster 7 December 2006.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
1Antonella Bozzi – LSC/Virgo meeting Amsterdam Ligo/Virgo Data Transfer Bulk data replication tools and plans for S6 data replication Antonella.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
D0 Taking Stock1 By Anil Kumar CD/CSS/DSG June 06, 2005.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
Portuguese Grid Infrastruture(s) Gonçalo Borges Jornadas LIP 2010 Braga, Janeiro 2010.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Virtual Server Server Self Service Center (S3C) JI July.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Open-E Data Storage Software (DSS V6)
Integrating Disk into Backup for Faster Restores
Experiences and Outlook Data Preservation and Long Term Analysis
Computing Infrastructure for DAQ, DM and SC
Design Unit 26 Design a small or home office network
Web Server Administration
Presentation transcript:

EGO Computing Center site report EGO - Via E. Amaldi S. Stefano a Macerata - Cascina (PI) | Stefano Cortese INFN Computing Workshop –

Alpha/OSF wks: 15 control 10 processing 2 servers x86/Linux nodes: 12 processing 4 servers 4 control LINUX farm nodes 16 nodes: 54 Gflop 2 servers 13 storage nodes 25 linux PCs On-line buffers 5TB OFF-LINE COMPUTING 60 LynxOS CPUs 10 OS9 CPUs 150 users Virgo-EGO computing areas ON-LINE PROCESSING/ IN-TIME domain ON-LINE PROCESSING/ IN-TIME domain Interferometer Real Time Domain Interferometer Real Time Domain Monitoring and Control Monitoring and Control DAQ 6 MB/s On-line Analysis 5-(300) Gflops Tape Backup 6TB > 6 MB/s Disk Storage 70TB Users computing and Data Access 34-(155) Mbps Bologna and Lyon Repositories OFFICE SERVICES 50 windows PCs OFFICE SERVICES 50 windows PCs INTERNET SERVICES Software integration testing, archiving and installation

Virgo-EGO LANs Firewall CheckPoint over Nokia 3 KM Virgo Interferometer network (> 50 switches) Offices UPS and generators Data Analysis network ( 7 switches) General windows PCs network ( 30 switches) WAN DMZ 34 Mbps

Storage HP_Compaq MA8000 Fiber Channel-To-SCSI 4 Terabytes 1 Week buffering Backup via Legato 6MB/s Accusys SCSI-To-SCSI 1 Terabyte Tape Library HP LTO ULTRIUM-1 6TB near-line Cataloguing MD5sum Virgo Data Redundant Stream Migration to mass storage Storage FARM: 13 nodes with 70 TB of net RAID5 space 25 Infortrend FC-to-IDE arrays 4 Fibrenetix Zero-d SCSI-to-IDE + some 3ware Mainly 250GB/7200rpm WD disks Everything with Linux RH9 and LVM 1.0.x

l Performances are good: 50/60MBytes/s over 1TB RAID5 set (single array, 5400/7200rpm disks) l The quality of the first releases of the products are very poor due mainly to firmware bugs or hardware tolerances that ultimately lead to hidden data corruption (meaning undetected by storage controller or operating system) l We developed a procedure for storage acceptance: requirement of minimal performances according to market survey and in demo testing Tenders are required in 2 lots, the first is for the acceptance test. The positive validation of the first lot is required for the acceptance of the second l Acceptance Test: Performances tested with “bonnie” Data integrity checked with a continuous benchmark that reads, writes and deletes data with 128bit MD5 verification after each operation Storage practices with IDE based arrays

l Data Integrity test: The data integrity test ends after the processing of about 30TB (after about 10days) giving confidence that the BER is less than 3x In our experience the errors may occur even after 1 week of processing and we rejected many configurations l The test needs to be repeated at each new firmware release installation, even if new features only are introduced l All this Is not enough: Many functions of the firmware may happen to be executed after the systems is running since a long time. That is the case of block remapping following bad block occurrences on the disks (this could only be tested using really bad disks) Therefore: The storage must be periodically monitored for data integrity The firmware must provide the on-line low level media verification that must executed periodically to avoid the double bad-blocks or bad- block+disk-failure cases Storage practices: data integrity

l IDE based storage systems at 5000€/TB are good for mass storage with fast access and high density compared to near-line disk-cache/tape systems but availability is not guaranteed at all times l They don’t offer the same level of reliability for critical tasks as more expensive disk based storage. Duplication or tape backup is still needed l Direct Attached Arrays are preferable respect to NAS storage to be able to run tests independently of the network We prefer also arrays connected via standard buses (e.g. SCSI or FC) rather than “on server” controllers to avoid intermixing OS/driver/array problems l LVM and automounter are required tools for mounting and serving about 100 file-systems (currently using amd, planning to pass to autofs on Linux) Storage Conclusions

On-line Computing Virgo detection channels are extracted from rawdata and processed to obtain the h-reconstructed signal where the Gravitational signal must be found 8 bi-processors Intel Xeon 2.66GHz 2 bi-processors Intel Xeon 2.0GHz The h-reconstructed signal ( KB/s) is fed to the computing farms for on- line search Small Scale Test System (2002) 16 bi-processors Compaq W GHz + PC800 RDRAM 2 front-ends 2 Standard gigabit ethernet LANs (internodes and storage)

On-line Computing: l Physical problem of coalescing binaries “in-time” detection was estimated by Virgo to require a 300GFlop system l Flat search with Matched filtering via FFT with templates of various length is very dependent on the amount of RAM available for storing the templates, so the naive sizing by CPU power is not enough l A benchmarking Virgo/EGO workgroup has been working since beginning of the year to arrive at more precise specifications (benchmark provided by Perugia group, tests performed by EGO)

Overall problem Opteron has the best speedup for SIMD problems where data are partitioned among processors: up to 60MB/s of template floats processed per CPU for the Virgo benchmark The Maximum RAM supported by the platform has an impact on the number of CPUs Overall Virgo problem for a space of templates (1.6 TB RAM) to be processed in 256s would require about 200 opteron with 8GB/CPU or 130 Itanium with 12GB/CPU Opteron has a higher performance per rack-unit Current tender is for 64 CPUs