DATA ACCESS and DATA MANAGEMENT CHALLENGES in CMS.

Slides:



Advertisements
Similar presentations
1 Jin Li Microsoft Research. Outline The Upcoming Video Tidal Wave Internet Infrastructure: Data Center/CDN/P2P P2P in Microsoft Locality aware P2P Conclusions.
Advertisements

DAKNET Presented By: rreema.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Copyright © Clifford Neuman and Dongho Kim - UNIVERSITY OF SOUTHERN CALIFORNIA - INFORMATION SCIENCES INSTITUTE Advanced Operating Systems Lecture.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
1. Maria Girone, CERN  Q WLCG Resource Utilization  Commissioning the HLT for data reprocessing and MC production  Preparing for Run II  Data.
NSF IRNC PI Meeting October 6, 2011 Julio Ibarra Center for Internet Augmented Research & Assessment Florida International University Americas Lightpaths.
1 BRUSSELS - 14 July 2003 Full Security Support in a heterogeneous mobile GRID testbed for wireless extensions to the.
1 08/Dec/01 S. F. Novaes Network in Brazil: Present and Future S. F. Novaes National Research Network Academic Network at São Paulo –Network Traffic –PingER.
Stefano Belforte INFN Trieste 1 CMS Simulation at Tier2 June 12, 2006 Simulation (Monte Carlo) Production for CMS Stefano Belforte WLCG-Tier2 workshop.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Cyberinfrastructure: An investment worth making Joe Breen University of Utah Center for High Performance Computing.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
Maria Girone, CERN CMS Experiment Status, Run II Plans, & Federated Requirements Maria Girone, CERN XrootD Workshop, January 27, 2015.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
Maria Girone, CERN  CMS in a High-Latency Environment  CMSSW I/O Optimizations for High Latency  CPU efficiency in a real world environment  HLT 
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
1 EUROPEAN COMMISSION Tempus JEP – – 2006 Supporting and facilitating active uptake to Information and Communication Technologies for University.
Fall 2005 Internet2 Member Meeting International Task Force Julio Ibarra, PI Heidi Alvarez, Co-PI Chip Cox, Co-PI John Silvester, Co-PI September 19, 2005.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Condor on Dedicated Clusters Peter Couvares and Derek Wright Computer Sciences Department University of Wisconsin-Madison
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Super Computing 2016 Salt Lake City, Utah Nov 15, 2016
AmLight-ExP: Collaboration at 100G
Digital Campus: Foundation Projects
TV Broadcasting What to look for Architecture TV Broadcasting Solution
Status of WLCG FCPPL project
Fast Data Transfer: Brazil Panel: Network Architecture for Globalization PRPv2 Workshop, UCSD Feb 21st, 2017 Michael Stanton RNP, Brazil.
Joint Genome Institute
Fan Engagement Solution
Computing Operations Roadmap
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Computing models, facilities, distributed computing
Innovation and RNP Panel “Innovation in the Balance” TERENA Networking Conference, June Maastricht, Netherlands Michael Stanton RNP – Brazil’s.
Christos Markou Institute of Nuclear Physics NCSR ‘Demokritos’
for the Offline and Computing groups
Joseph JaJa, Mike Smorul, and Sangchul Song
Design and Demo of AtlanticWave/SDX
LHC DATA ANALYSIS INFN (LNL – PADOVA)
Luca dell’Agnello INFN-CNAF
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
UK Status and Plans Scientific Computing Forum 27th Oct 2017
Thoughts on Computing Upgrade Activities
SENSE: SDN for End-to-end Networked Science at the Exascale
Network Requirements Javier Orellana
Multi-Processing in High Performance Computer Architecture:
K. Schauerhammer, K. Ullmann (DFN)
Introduction to client/server architecture
Storage Virtualization
Multi-Processing in High Performance Computer Architecture:
NSF cloud Chameleon: Phase 2 Networking
Introduction to Operating Systems
Any Data, Anytime, Anywhere
EE 122: Lecture 22 (Overlay Networks)
The UltraLight Program
SANDIE: Optimizing NDN for Data Intensive Science
Presentation transcript:

DATA ACCESS and DATA MANAGEMENT CHALLENGES in CMS

 CMS produces ~20PB of raw and derived data per year  An average replication factor of ~3  70 Computing sites that are globally distributed  How to deliver samples to 150k processor cores as directed by the experiment centrally and thousands of scientists

 The network capacity itself is keeping pace (just barely) due to the availability of 100Gb/s links  However we have a factor of 100 between are best and worst connected sites  Our ability to drive the network efficiently is still an issue, we use a lot of hardware to fill the pipes

Upload rate: 27 Gbps; 20Gbps to CNAF (Italy) Alone ● By Spring 2015: 12 – 40 Gbps Downloads were Routine to US CMS Tier2 Sites with 100G Links Downloading Terabyte Datasets to Tier3s desktop/laptops is being explored ● 30 US CMS university based Tier2s have moved to ~100G now routine The move to 100G is timely and matches current needs, also at Tier2s. Backbones should continue to advance to meet the needs during Run2. Harvey Newman

OpenWave: First 100G Link to Latin America in Connecting LSST 10G  An “Alien Wave” at 100G on the Undersea Cable  Precedent-setting access to the frequency spectrum by the academic community  Sao Paulo-Rio-Fortaleza -St. Croix-Miami backbone  Scheduled to start soon  100G extensions by RNP in Rio and ANSP in Sao Paulo  Will be extended to Chile at 100G then N X 100G  Will be heavily used by LSST into the 2030s AmLight (US NSF) with RNP, ANSP 10G 100G 10G J. Ibarra, AmLight February 2015  Using Padtec (BR) 100G equipment. Demonstrations with the HEP team (Caltech et al) at SC2013 and 2014 Total Capacity for Next Two Years: 140G Harvey Newman

 We have spent the last several years trying to allow the processing and storage services to be more independent  Disk is expensive and normally has move IO capability than the amount of local processing services  Before this there was a lot of worry about the balance of CPU and storage  CPU can be scheduled more dynamically  CPU can be used opportunistically

Data Federation in Run II  We validated small scale use of non- local data access in the summer  Fall-back when analysis jobs do not find data  Very good feedback by users  After summer scale tests were performed in Europe and the US  20% of jobs were able to access data over the wide area (60k files/day, O(100TB)/day)  Production system for Run2 enabling  Interactive access  Fail over protection  The ability to share production workflows 7 Any Data Anytime Anywhere has been a primary focus area

Maria Girone, CHEP 2015  xxx 8 Aggregate bandwidth > 2GB/s Each site has delivered PBs over the last year CMS

 Data Federation is not a content delivery network (CDN)  It has only basic network awareness  Integration of more intelligent caching and intermediate storage  We see interesting opportunities in development of advanced data management that begins to close the gap between data federation and CDN  End goal would be to care a lot less about the actual location of the data  Looking forward we would like to investigate Named Data Networks where more of the data management is integrated with the network itself