1 R. Voicu 1, I. Legrand 1, H. Newman 1 2 C.Grigoras 1 California Institute of Technology 2 CERN CHEP 2010 Taipei, October 21 st, 2010 End to End Storage.

Slides:



Advertisements
Similar presentations
Middleware Support for RDMA-based Data Transfer in Cloud Computing Yufei Ren, Tan Li, Dantong Yu, Shudong Jin, Thomas Robertazzi Department of Electrical.
Advertisements

MONITORING WITH MONALISA Costin Grigoras. M ONITORING WITH M ON ALISA What is MonALISA ? MonALISA communication architecture Monitoring modules ApMon.
October 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
1 I/O Management in Representative Operating Systems.
Traffic shaping with OVS and SDN Ramiro Voicu Caltech LHCOPN/LHCONE, Berkeley, June
Institute of Computer Science AGH Performance Monitoring of Java Web Service-based Applications Włodzimierz Funika, Piotr Handzlik Lechosław Trębacz Institute.
Grid Monitoring By Zoran Obradovic CSE-510 October 2007.
ALICE DATA ACCESS MODEL Outline ALICE data access model - PtP Network Workshop 2  ALICE data model  Some figures.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
© 2008 Cisco Systems, Inc. All rights reserved.CIPT1 v6.0—2-1 Administering Cisco Unified Communications Manager Understanding Cisco Unified Communications.
Globus Striped GridFTP Framework and Server Raj Kettimuthu, ANL and U. Chicago.
ALICE data access WLCG data WG revival 4 October 2013.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems.
Online Monitoring with MonALISA Dan Protopopescu Glasgow, UK Dan Protopopescu Glasgow, UK.
◦ What is an Operating System? What is an Operating System? ◦ Operating System Objectives Operating System Objectives ◦ Services Provided by the Operating.
Module 10: Monitoring ISA Server Overview Monitoring Overview Configuring Alerts Configuring Session Monitoring Configuring Logging Configuring.
Open Science Grid The OSG Accounting System: GRATIA by Philippe Canal (FNAL) & Matteo Melani (SLAC) Mumbai, India CHEP2006.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
1 Ramiro Voicu, Iosif Legrand, Harvey Newman, Artur Barczyk, Costin Grigoras, Ciprian Dobre, Alexandru Costan, Azher Mughal, Sandor Rozsa Monitoring and.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
Monitoring, Accounting and Automated Decision Support for the ALICE Experiment Based on the MonALISA Framework.
February 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology February 2006 February 2006 An Agent Based, Dynamic Service System to Monitor,
1 Iosif Legrand, Harvey Newman, Ramiro Voicu, Costin Grigoras, Catalin Cirstoiu, Ciprian Dobre An Agent Based, Dynamic Service System to Monitor, Control.
Swapping to Remote Memory over InfiniBand: An Approach using a High Performance Network Block Device Shuang LiangRanjit NoronhaDhabaleswar K. Panda IEEE.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
N EWS OF M ON ALISA SITE MONITORING
Servicii distribuite Alocarea dinamică a resurselor de reea pentru transferuri de date de mare viteză folosind servicii distribuite Distributed Services.
Site operations Outline Central services VoBox services Monitoring Storage and networking 4/8/20142ALICE-USA Review - Site Operations.
Operating Systems David Goldschmidt, Ph.D. Computer Science The College of Saint Rose CIS 432.
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Overview of ALICE monitoring Catalin Cirstoiu, Pablo Saiz, Latchezar Betev 23/03/2007 System Analysis Working Group.
Local Monitoring at SARA Ron Trompert SARA. Ganglia Monitors nodes for Load Memory usage Network activity Disk usage Monitors running jobs.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
Monitoring with MonALISA Costin Grigoras. What is MonALISA ?  Caltech project started in 2002
1 MonALISA Team Iosif Legrand, Harvey Newman, Ramiro Voicu, Costin Grigoras, Ciprian Dobre, Alexandru Costan MonALISA capabilities for the LHCOPN LHCOPN.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
April 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
ALICE DATA ACCESS MODEL Outline 05/13/2014 ALICE Data Access Model 2  ALICE data access model  Infrastructure and SE monitoring.
+ AliEn site services and monitoring Miguel Martinez Pedreira.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
October 2006 Iosif Legrand 1 Iosif Legrand California Institute of Technology An Agent Based, Dynamic Service System to Monitor, Control and Optimize Distributed.
03/09/2007http://pcalimonitor.cern.ch/1 Monitoring in ALICE Costin Grigoras 03/09/2007 WLCG Meeting, CHEP.
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR F2F Monitoring at CERN Miguel Coelho dos Santos.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
MONITORING WITH MONALISA Costin Grigoras. M ON ALISA COMMUNICATION ARCHITECTURE MonALISA software components and the connections between them Data consumers.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
MONALISA MONITORING AND CONTROL Costin Grigoras. O UTLINE MonALISA services and clients Usage in ALICE Online SE discovery mechanism Data management 3.
PHD Virtual Technologies “Reader’s Choice” Preferred product.
Federating Data in the ALICE Experiment
Jean-Philippe Baud, IT-GD, CERN November 2007
ECSAC- August 2009, Veli Losink California Institute of Technology
California Institute of Technology
ALICE Monitoring
PROOF – Parallel ROOT Facility
Sergio Fantinel, INFN LNL/PD
Integration of Network Services Interface version 2 with the JUNOS Space SDK
Storage elements discovery
Ákos Frohner EGEE'08 September 2008
Enabling High Speed Data Transfer in High Energy Physics
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Presentation transcript:

1 R. Voicu 1, I. Legrand 1, H. Newman 1 2 C.Grigoras 1 California Institute of Technology 2 CERN CHEP 2010 Taipei, October 21 st, 2010 End to End Storage Performance Measurements

2 Introduction  Understanding the end to end performance of storage systems to move large amounts of data over the wide area network is critical for data management planning and very useful for debugging such complex and heterogeneous systems  In general, each storage system has a set of data servers. The number of data servers varies from a few to hundreds and is very much dependent on the type of storage technology used by a site.  The entire system is highly heterogeneous : network, storage and transfer application wise  An End to End monitoring should bring a common view and should help in debugging and understanding data transfer performance in a transparent manner

3 Data moving patterns Network infrastructure (LAN + WAN) Site A Site B Storage cluster Data transfer gateway(s) Data transfer gateway(s) Data flow

4 Network infrastructure (LAN + WAN) Site ASite B Storage cluster Data flow Data moving patterns (2)

5 Monitoring metrics Network infrastructure (LAN + WAN) Site B Storage cluster Data transfer gateway(s) Data transfer gateway(s) App Monitoring CPU usage, Memory Disk to Network Queue monitoring Full HW&OS Monitoring CPU usage, Memory, Load, Eth traffic, etc Disk IO Load Net Monitoring Net interfaces traffic Error countersNetflow/SFlow

6 Motivation  The MonALISA monitoring system includes:  Local host monitoring (CPU, memory, network traffic, Disk I/O, processes and sockets in each state, LM sensors, APC UPSs), log files tailing  SNMP generic & specific modules  Condor, PBS, LSF and SGE (accounting & host monitoring), Ganglia  Ping, tracepath, traceroute, pathload and other network-related measurements  TL1, Network devices, Ciena, Optical switches  Calling external applications/scripts that return as output the values  XDR-formatted UDP messages (such as ApMon)

7 Motivation(cont) Monitoring architecture in ALICE 7 Long History DB LCG Tools ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn Job Agent MonALISA LCG Site ApMon AliEn CE ApMon AliEn SE ApMon Cluster Monitor ApMon AliEn TQ ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn Job Agent ApMon AliEn CE ApMon AliEn SE ApMon Cluster Monitor ApMon AliEn IS ApMon AliEn Optimizers ApMon AliEn Brokers ApMon MySQL Servers ApMon CastorGrid Scripts ApMon API Services MonaLisaRepository Aggregated Data rss vsz cpu time run time job slots free space nr. of files open files Queued JobAgents cpu ksi2k job status disk used processes load net In/out jobs status sockets migrated mbytes active sessions MyProxy status Alerts Actions Orchestrated Data transfer tool Orchestrated Orchestrated

FDT – Fast Data Transfer  FDT is an open source application, developed at Caltech, for efficient data transfers  Easy to use: similar syntax with SCP, iperf/netperf  Written in java and runs on all major platforms.  Single.jar file (~800 KB)  Based on an asynchronous, multithreaded system  Uses the New I/O (NIO) interface and is able to:  stream continuously a list of files  use independent threads to read and write on each physical device  transfer data in parallel on multiple TCP streams, when necessary  use appropriate buffer size for disk IO and networking  resume a file transfer session

FDT - Architecture Pool of buffers Kernel Space Data Transfer Sockets / Channels Independent threads per device Restore the files from buffers Control connection / authorization Pool of buffers Kernel Space

FDT features  User defined loadable modules for Pre and Post Processing to provide support for dedicated Mass Storage system, compression, dynamic circuit setup, …  Pluggable file systems “providers” (e.g. non-POSIX FS)  Dynamic bandwidth capping (can be controlled by MonALISA)  Different transport strategies:  blocking (1 thread per channel)  non-blocking (selector + pool of threads)  On the fly MD5 checksum on the reader side  Configurable number of streams and threads per physical device (useful for distributed FS)  Automatic updates  Can be used as network testing tool (/dev/zero → /dev/null memory transfers, or –nettest flag)

FDT Throughput tests – 1 Stream

12 Active End to End Available Bandwidth between all the ALICE grid sites

Active End to End Available Bandwidth between all the ALICE grid sites (2) 1 Gbps network card Newer kernel Tuned TCP Buffers 100 Mbps network card Default kernels Default TCP Buffers Different trends = different kernels

14 CPU and Disk I/O performance metrics

15 CPU and Disk I/O performance metrics 8 Fast SAS Disks RAID6 512MB Raid controller Clear correlation Expected behavior

16 CPU and Disk I/O performance metrics <8 Mbytes/s RW Speed CPU Idle > 85% CPU IOWait ~ 10-11% IO Utilization 100% !! Same I/O hardware: 8 Fast SAS Disks RAID6 512MB Raid controller

17  MonALISA provides a wide set of monitoring modules  Full host system monitoring: CPU usage, System load, Disk IO/Load, Memory, Swap, etc  Network infrastructure monitoring and topology  Application monitoring (via ApMon)  Local and global triggers and alarms  Powerful correlation framework ; a few lines of config file to send an alert ( , SMS, IM, etc) or take an action (e.g. restarting a service)  The security and controlling infrastructure is built-in.  Orchestrated end to end tests (bandwidth and disk) are scheduled using secure channels  End to end network bandwidth tests (memory to memory) Current status

18 Future plans  Add per process/task I/O disk statistics  Add local disk performance tests (only when the disk is idle) to get a baseline  Extend the End to End tests to the entire chain:  Site A Disk => Network => Site B Disk  Check if the network and disk to disk baseline match  Raise alarms in case of problems  Disk IO Utilization to R/W I/O ratio below a certain threshold  Current performance below the established baseline  Integrate with the network topology measurements (already used to choose best SE, based on RTT)

19 Q&A