Www.ci.anl.gov www.ci.uchicago.edu Brainstorming Thin-provisioned Tier 3s.

Slides:



Advertisements
Similar presentations
Tom Byrne, 12 th November 2014 Ceph – status update and xrootd testing Alastair Dewhurst, Tom Byrne 1.
Advertisements

Simple Git Steve Pieper. Topics Git considerations and Slicer Git as if it were svn Git the way it is meant to be.
A comparison between xen and kvm Andrea Chierici Riccardo Veraldi INFN-CNAF.
Patterns & practices Symposium 2013 Introducing Git version control into your team Mark
Dynamics AX Technical Overview Application Architecture Dynamics AX Technical Overview.
Tier-1 experience with provisioning virtualised worker nodes on demand Andrew Lahiff, Ian Collier STFC Rutherford Appleton Laboratory, Harwell Oxford,
Version Control with git. Version Control Version control is a system that records changes to a file or set of files over time so that you can recall.
1 CSE 390 “Lecture 11” Version control with Git slides created by Ruth Anderson, images from
Git for Version Control These slides are heavily based on slides created by Ruth Anderson for CSE 390a. Thanks, Ruth! images taken from
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Virtual Servers Harry Dickens Arkansas Public School Resource Center Harry Dickens Arkansas Public School Resource Center.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Block1 Wrapping Your Nugget Around Distributed Processing.
CERN IT Department CH-1211 Genève 23 Switzerland t Experiences running a production Puppet Ben Jones HEPiX Bologna Spring.
INFSO-RI Module 01 ETICS Overview Etics Online Tutorial Marian ŻUREK Baltic Grid II Summer School Vilnius, 2-3 July 2009.
Configuration Management with Cobbler and Puppet Kashif Mohammad University of Oxford.
How to create DNS rule that allow internal network clients DNS access Right click on Firewall Policy ->New- >Access Rule Right click on Firewall.
Condor and DRBL Bruno Gonçalves & Stefan Boettcher Emory University.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Experience with the Thumper Wei Yang Stanford Linear Accelerator Center May 27-28, 2008 US ATLAS Tier 2/3 workshop University of Michigan, Ann Arbor.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Arne Wiebalck -- VM Performance: I/O
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Tim Bell 04/07/2013 Intel Openlab Briefing2.
PerfSONAR Update Shawn McKee/University of Michigan LHCONE/LHCOPN Meeting Cambridge, UK February 9 th, 2015.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
The VL-e Proof of Concept Environment & The VL-e PoC Installer Jan Just Keijser System Integrator P4 Team NIKHEF.
SECTION 1: CODE REASONING + VERSION CONTROL slides borrowed and adapted from Alex Mariakis and CSE 390a Justin Bare & Deric Pang.
HTCondor Private Cloud Integration Andrew Lahiff STFC Rutherford Appleton Laboratory European HTCondor Site Admins Meeting 2014.
1 Update at RAL and in the Quattor community Ian Collier - RAL Tier1 HEPiX FAll 2010, Cornell.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
1 Cloud Services Requirements and Challenges of Large International User Groups Laurence Field IT/SDC 2/12/2014.
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Introduce Caching Technologies using Xrootd Wei Yang 10/28/14ATLAS TIM 2014 Univ. Chicago1.
INFSO-RI Enabling Grids for E-sciencE Distributed Version Control Systems Ákos Frohner CERN, IT-DM.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
CERN AI Config Management 16/07/15 AI for INFN visit2 Overview for INFN visit.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Cloud Installation & Configuration Management. Outline  Definitions  Tools, “Comparison”  References.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
The Great Migration: From Pacman to RPMs Alain Roy OSG Software Coordinator.
Platform & Engineering Services CERN IT Department CH-1211 Geneva 23 Switzerland t PES Agile Infrastructure Project Overview : Status and.
What is Cloud Computing 1. Cloud computing is a service that helps you to perform the tasks over the Internet. The users can access resources as they.
SCDB Update Michel Jouvin LAL, Orsay March 17, 2010 Quattor Workshop, Thessaloniki.
Out of the basement, into the cloud: Extending the Tier 3 Lincoln Bryant Computation and Enrico Fermi Institutes.
Efi.uchicago.edu ci.uchicago.edu Caching FAX accesses Ilija Vukotic ADC TIM - Chicago October 28, 2014.
BY: SALMAN 1.
Site Administration Tools: Ansible
BY: SALMAN.
Virtualisation for NA49/NA61
Dag Toppe Larsen UiB/CERN CERN,
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Dag Toppe Larsen UiB/CERN CERN,
Introduction to CVMFS A way to distribute HEP software on cloud
Virtualisation for NA49/NA61
Modern Version Control with Git
Spacewalk and Koji at Fermilab
Building Grids with Condor
Integration of Singularity With Makeflow
Version Control with Git and GitHub
Presentation transcript:

Brainstorming Thin-provisioned Tier 3s

2 Masterless Puppet for configuration Puppet is fairly straightforward to set up in an “masterless” mode. – Simply distribute Puppet rules to nodes via VCS (git, mercurial, SVN, whatever) – Some more reading: distributed-version-control Common software installed via git repository, applied locally. essentially: – $ yum install puppet – $ git clone – $ vim Tier3/init.pp – $ puppet apply Tier3/init.pp Voila?

3 Advantages of Puppet-based T3 Large knowledge base to draw upon – MWT2 and UC ATLAS Tier 3 are managed via Puppet – Happy to share these puppet rules with the community Will be available via Github repo – Contributions welcome Easy to cherrypick certain features We are planning to build a few T3 interactive/analysis boxes and will lay the groundwork for this

4 Kernel bCache Transparent kernel block layer cache. – Problem: o SSDs = Fast but low capacity o Spinning disks = Slow but high capacity – Solution: Put SSDs in front of RAID array to cache reads/writes o Totally transparent to the filesystem o Write-through, write-back, and write-around modes available o Save $$$ by moving high-end hardware RAID features into kernel space Considered stable in kernel – Might be too aggressive for some conservative sites – At MWT2 we routinely use kernel 3.11 on interactive nodes.

5 FAX + XRootD Cache Confgure T3 nodes with a large disk array, optionally use kernel bcache for performance boost. Use FAX to read data off of the grid Cache data locally with XRootD sub-file level caching – New feature in XRootD 4

6 ATLAS Connect for batch T3 puppet rules are pre-configured to install Condor & flock to ATLAS Connect Need to figure out authentication – Should aim to be as lightweight as possible – X509 is greatly not-preferable, as a host cert/key pair has to be generated for each site. – Shared password viable? Seems reasonable if password is stored in a private repository on GitHub

7 Extra slide: Other Methods of distribution Couple of different approaches we could take – Red Hat Spacewalk o Build a SL image for ATLAS Tier 3 nodes o Essentially deprecated at this point. RH is focused on cloud deployments, i.e., Openstack + Puppet/Chef/Ansible/Configuration management tool du jour – ATLAS T3 RPMs o Wrap up some common configs and RPMize them. i.e., – ATLAS Connect condor config?  condor-atlas-connect.rpm – Preconfigured XRootD Cache w/ FAX upstream redirection?  xrootd-atlas-fax-cache.rpm o More flexible but probably overall more work for little gain – Centrally managed Puppet o Puppet daemon is advantageous since it runs every 15 minutes, easy to push out updates o Easy enough to write a cron to do similar task