Presentation is loading. Please wait.

Presentation is loading. Please wait.

14th October 2014Graduate Lectures1 Oxford University Particle Physics Unix Overview Sean Brisbane Particle Physics Systems Administrator Room 661 Tel.

Similar presentations


Presentation on theme: "14th October 2014Graduate Lectures1 Oxford University Particle Physics Unix Overview Sean Brisbane Particle Physics Systems Administrator Room 661 Tel."— Presentation transcript:

1 14th October 2014Graduate Lectures1 Oxford University Particle Physics Unix Overview Sean Brisbane Particle Physics Systems Administrator Room 661 Tel 73389 s.brisbane1@physics.ox.ac.uk

2 14th October 2014Graduate Lectures2 l Strategy l Local Cluster Overview l Connecting to it l Grid Cluster l Computer Rooms l How to get help

3 14th October 2014Graduate Lectures3 Particle Physics Strategy The Server / Desktop Divide Win 7 PC Linux Desktop Desktops Servers General Purpose Unix Server Group DAQ Systems Linux Worker nodes Web Server Linux File Servers Win 7 PC Ubuntu PC Approx 200 Desktop PC’s with Exceed, putty or ssh/X windows used to access PP Linux systems Virtual Machine Host NIS Server torque Server

4 Physics fileservers and clients 14th October 2014Graduate Lectures4 Windows server Central Linux file- server PP file- server Storage system: ClientWindowsCentral Ubuntu PP Linux Recommended storage H:\ drive/home folder/home and /data folders Windows storage“H:\” drive or “Y:\home” /physics/home PP StorageY:/LinuxUsers/pplinux /data/home /data/home, /data/experiment Central LinuxY:/LinuxUsers/home/ particle /network/home/particle

5 14th October 2014Graduate Lectures5 Particle Physics Linux l Unix Team (Room 661): n Pete Gronbech - Senior Systems Manager and GridPP Project Manager n Ewan MacMahon – Grid Systems Administrator n Kashif Mohammad – Grid and Local Support n Sean Brisbane – Local Server and User Support l General purpose interactive Linux based systems for code development, short tests and access to Linux based office applications. These are accessed remotely. l Batch queues are provided for longer and intensive jobs. Provisioned to meet peak demand and give a fast turnaround for final analysis. l Systems run Scientific Linux (SL) which is a free Red Hat Enterprise based distribution. l The Grid & CERN have migrated to SL6. The majority of the local cluster is also on SL6, but some legacy SL5 systems are provided for those that need them. l We will be able to offer you the most help running your code on the newer SL6. Some experimental software frameworks still require SL5.

6 14th October 2014Graduate Lectures6 Current Clusters l Particle Physics Local Batch cluster l Oxfords Tier 2 Grid cluster

7 14th October 2014 PP Linux Batch Farm pplxwn15 Scientific Linux 6 pplxint8 pplxint9 8 * Intel 5420 cores Interactive login nodes pplxwn16 8 * Intel 5420 cores 7Graduate Lectures pplxwn31 pplxwn32 pplxwnnn pplxwn38 pplxwn41 pplxwnnn pplxwn60 16 * Intel cores 16 * Intel 2650 cores 12 * Intel 5650 cores Users log in to the interactive nodes pplxint8 & 9, the home directories and all the data disks (/home area or /data/group ) are shared across the cluster and visible on the interactive machines and all the batch system worker nodes. Approximately 300 cores (430 incl. JAI/LWFA), each with 4GB of RAM memory. The /home area is where you should keep your important text files such as source code, papers and thesis The /data/ area is where you should put your big reproducible input and output data pplxwnnn 12 * Intel 5650 cores pplxwn59 jailxwn01 64 * AMD cores 16 * Intel cores jailxwn02 64 * AMD cores

8 14th October 2014 PP Linux Batch Farm Scientific Linux 5 pplxint5 Interactive login nodes pplxwn23 16 * AMD 6128 cores pplxwn24 16 * AMD 6128 cores 8Graduate Lectures Legacy SL5 jobs supported by smaller selection of worker nodes. Currently eight servers with 16 cores each with 4GB of RAM memory per core. All of your files area available from SL5 and 6, but the software environment will be different and therefore your code may not run if compiled for the other operating system. pplxwn30 16 * AMD 6128 cores pplxwnnn 16 * AMD 6128 cores pplxwnnn 16 * AMD 6128 cores pplxint6

9 14th October 2014 PP Linux Batch Farm Data Storage pplxfsn 40TB pplxfsn 40TB Data Areas pplxfsn 19TB 9 Graduate Lectures NFS Servers Home areas Data Areas NFS is used to export data to the smaller experimental groups, where the partition size is less than the total size of a server. The data areas are too big to be backed up. The servers have dual redundant PSUs, RAID 6 and are running on uninterruptible powers supplies. This safeguards against hardware failures, but does not help if you delete files. The home areas are backed up by two different systems nightly. The Oxford ITS HFS service and a local back up system. If you delete a file tell us a soon as you can when you deleted it and it’s full name. The latest nightly backup of any lost or deleted files from your home directory is available at the read-only location /data/homebackup/{username} The home areas are quota’d but if you require more space ask us. Store your thesis on /home NOT /data. pplxfsn 30TB Data Areas

10 Particle Physics Computing Lustre MDSLustre OSS02Lustre OSS03 18TB 44TB SL6 Node SL5 Node pplxint8 Lustre OSS01 44TB Lustre OSS04 df -h /data/atlas Filesystem Size Used Avail Use% Mounted on /lustre/atlas25/atlas 366T 199T 150T 58% /data/atlas df -h /data/lhcb Filesystem Size Used Avail Use% Mounted on /lhcb25 118T 79T 34T 71% /data/lhcb25 14th October 201410 Graduate Lectures The Lustre file system is used to group multiple file servers together to provide extremely large continuous file spaces. This is used for the Atlas and LHCb groups. pplxint5

11 14th October 2014Graduate Lectures11

12 14th October 2014Graduate Lectures12 Strong Passwords etc l Use a strong password not open to dictionary attack! n fred123 – No good n Uaspnotda!09 – Much better l More convenient* to use ssh with a passphrased key stored on your desktop. n Once set up

13 14th October 2014Graduate Lectures13 Connecting with PuTTY Question: How many of you are using Windows? & Linux? On the desktop Demo 1. Plain ssh terminal connection 1. From ‘outside of physics’ 2. From Office (no password) 2. ssh with X windows tunnelled to passive exceed 3. ssh, X windows tunnel, passive exceed, KDE Session 4. Password-less access from ‘outside physics’ 1. See backup slides http://www2.physics.ox.ac.uk/it-services/ppunix/ppunix-cluster http://www.howtoforge.com/ssh_key_based_logins_putty

14 14th October 2014Graduate Lectures14

15 14th October 2014Graduate Lectures15 SouthGrid Member Institutions l Oxford l RAL PPD l Cambridge l Birmingham l Bristol l Sussex l JET at Culham

16 Current capacity l Compute Servers n Twin and twin squared nodes –1770 CPU cores l Storage n Total of ~1300TB n The servers have between 12 and 36 disks, the more recent ones are 4TB capacity each. These use hardware RAID and UPS to provide resilience. 14th October 2014Graduate Lectures16

17 14th October 2014Graduate Lectures17 Get a Grid Certificate Must remember to use the same PC to request and retrieve the Grid Certificate. The new UKCA page http://www.ngs.ac.uk/ukca uses a JAVA based CERT WIZARD You will then need to contact central Oxford IT. They will need to see you, with your university card, to approve your request: To: help@it.ox.ac.uk Dear Stuart Robeson and Jackie Hewitt, I Please let me know a good time to come over to Banbury road IT office for you to approve my grid certificate request. Thanks.

18 When you have your grid certificate… 14th October 2014 Graduate Lectures18 Save to a filename in your home directory on the Linux systems, eg: Y:\Linuxusers\particle\home\{username}\mycert.p12 Log in to pplxint9 and run mkdir.globus chmod 700.globus cd.globus openssl pkcs12 -in../mycert.p12 -clcerts -nokeys -out usercert.pem openssl pkcs12 -in../mycert.p12 -nocerts -out userkey.pem chmod 400 userkey.pem chmod 444 usercert.pem

19 Now Join a VO l This is the Virtual Organisation such as “Atlas”, so: n You are allowed to submit jobs using the infrastructure of the experiment n Access data for the experiment l Speak to your colleagues on the experiment about this. It is a different process for every experiment! 14th October 2014Graduate Lectures19

20 Joining a VO l Your grid certificate identifies you to the grid as an individual user, but it's not enough on its own to allow you to run jobs; you also need to join a Virtual Organisation (VO). l These are essentially just user groups, typically one per experiment, and individual grid sites can choose to support (or not) work by users of a particular VO. l Most sites support the four LHC VOs, fewer support the smaller experiments. l The sign-up procedures vary from VO to VO, UK ones typically require a manual approval step, LHC ones require an active CERN account. l For anyone that's interested in using the grid, but is not working on an experiment with an existing VO, we have a local VO we can use to get you started. 14th October 2014Graduate Lectures20

21 When that’s done l Test your grid certificate: > voms-proxy-init –voms lhcb.cern.ch Enter GRID pass phrase: Your identity: /C=UK/O=eScience/OU=Oxford/L=OeSC/CN=j bloggs Creating temporary proxy..................................... Done l Consult the documentation provided by your experiment for ‘their’ way to submit and manage grid jobs 14th October 2014Graduate Lectures21

22 14th October 2014Graduate Lectures22 Two Computer Rooms provide excellent infrastructure for the future The New Computer room built at Begbroke Science Park jointly for the Oxford Super Computer and the Physics department, provides space for 55 (11KW) computer racks. 22 of which will be for Physics. Up to a third of these can be used for the Tier 2 centre. This £1.5M project was funded by SRIF and a contribution of ~£200K from Oxford Physics. The room was ready in December 2007. Oxford Tier 2 Grid cluster was moved there during spring 2008. All new Physics High Performance Clusters will be installed here.

23 14th October 2014Graduate Lectures23 Local Oxford DWB Physics Infrastructure Computer Room Completely separate from the Begbroke Science park a computer room with 100KW cooling and >200KW power has been built. ~£150K Oxford Physics money. Local Physics department Infrastructure computer room. Completed September 2007. This allowed local computer rooms to be refurbished as offices again and racks that were in unsuitable locations to be re housed.

24 Cold aisle containment 24 14th October 2014Graduate Lectures

25 Other resources (for free) l Oxford Advanced Research Computing n A shared cluster of CPU nodes, “just” like the local cluster here n GPU nodes –Faster for ‘fitting’, toy studies and MC generation –*IFF* code is written in a way that supports them n Moderate disk space allowance per experiment (<5TB) n http://www.arc.ox.ac.uk/content/getting-started http://www.arc.ox.ac.uk/content/getting-started l Emerald n Huge farm of GPUs n http://www.cfi.ses.ac.uk/emerald/ http://www.cfi.ses.ac.uk/emerald/ l Both needs a separate account and project n Come talk to us in RM 661 14th October 2014Graduate Lectures25

26 14th October 2014Graduate Lectures26 The end of the overview l Now more details of use of the clusters l Help Pages n http://www.physics.ox.ac.uk/it/unix/default.htm http://www.physics.ox.ac.uk/it/unix/default.htm n http://www2.physics.ox.ac.uk/research/particle- physics/particle-physics-computer-support http://www2.physics.ox.ac.uk/research/particle- physics/particle-physics-computer-support l ARC n http://www.arc.ox.ac.uk/content/getting-started http://www.arc.ox.ac.uk/content/getting-started l Email n pp_unix_admin@physics.ox.ac.uk

27 BACKUP 14th October 2014Graduate Lectures27

28 Puttygen to create an ssh key on Windows (previous slide point #4) 14th October 2014Graduate Lectures28 Paste this into ~/.ssh/authorized_keys on pplxint Enter a secure passphrase then : - Enter a strong passphrase - Save the private parts of the key to a subdirectory of your local drive.

29 Pageant l Run Pageant once after login l Right-click on the pageant symbol and and “Add key” for your Private (windows ssh key) 14th October 2014Graduate Lectures29

30 14th October 2014Graduate Lectures30 Network l Gigabit JANET connection to campus July 2005. l Second JANET gigabit connection Sept 2007. l JANET campus connection upgraded to dual 10 gigabit links August 2009 l Gigabit Juniper firewall manages internal and external Physics networks. l 10Gb/s network links installed between Tier-2 and Tier-3 clusters in 2012. l Physics-wide wireless network. Installed in DWB public rooms, Martin Wood, AOPP and Theory. New firewall provides routing and security for this network.

31 14th October 2014Graduate Lectures31 Network Access Campus Backbone Router Super Janet 4 2* 10Gb/s with Janet 6 OUCS Firewall depts Physics Firewall Physics Backbone Router 1Gb/s 10Gb/s 1Gb/s 10Gb/s Backbone Edge Router depts 100Mb/s 1Gb/s depts 100Mb/s Backbone Edge Router 10Gb/s

32 14th October 201432 Physics Backbone desktop Server switch Physics Firewall Physics Backbone Switch Dell 8024F 10Gb/s 1Gb/s Particle Physics Dell 8024F desktop 1Gb/s Clarendon Lab Dell 8024F 10Gb/s Win 2k Server Astro Dell 8024F 10Gb/s 1Gb/s Theory Dell 8024F 10Gb/s Atmos Dell 8024F 10Gb/s Server Switch S4810 10Gb/s Linux Server 10Gb/s Linux Server 10Gb/s Super FRODO Frodo 10Gb/s 1Gb/s Graduate Lectures


Download ppt "14th October 2014Graduate Lectures1 Oxford University Particle Physics Unix Overview Sean Brisbane Particle Physics Systems Administrator Room 661 Tel."

Similar presentations


Ads by Google