Presentation is loading. Please wait.

Presentation is loading. Please wait.

R. Pordes, I Brazilian LHC Computing Workshop 1 What is Open Science Grid?  High Throughput Distributed Facility  Shared opportunistic access to existing.

Similar presentations


Presentation on theme: "R. Pordes, I Brazilian LHC Computing Workshop 1 What is Open Science Grid?  High Throughput Distributed Facility  Shared opportunistic access to existing."— Presentation transcript:

1 R. Pordes, I Brazilian LHC Computing Workshop 1 What is Open Science Grid?  High Throughput Distributed Facility  Shared opportunistic access to existing clusters, storage and networks.  Owner controlled resources and usage policies.  Supports Science  Funded by NSF and DOE projects.  Common technologies & cyber-infrastructure.  Open and Heterogeneous  Research groups transitioning from & extending (legacy) systems to Grids:  Experiments developing new systems.  Application Computer Scientists  Real life use of technology, integration, operation.

2 R. Pordes, I Brazilian LHC Computing Workshop 2 Who is OSG: a Consortium US DOE HENP Laboratory Facilities + Universities (US) LHC Collaborations + offshore sites LIGO Condor Project Running HENP Experiments - CDF, D0, STAR… Globus/CDIGS LBNL SDM Collaboration of users, developers, grid technologists, facility administrators. Training & help for administrators and users

3 R. Pordes, I Brazilian LHC Computing Workshop 3 OSG 1 day last week: 50 Clusters : used locally as well as through the grid 5 Large disk or tape stores 23 VOs >2000 jobs running through Grid; Bioinformatics Routed from Local UWisconsin Campus Grid 2000 running jobs 500 waiting jobs LHC Run II

4 R. Pordes, I Brazilian LHC Computing Workshop 4 Broad Engagement

5 R. Pordes, I Brazilian LHC Computing Workshop 5 The OSG World: Partnerships Campus Grids:  GRid Of IoWa,  Grid Laboratory Of Wisconsin,  Crimson Grid,  Texas Advanced Computer Center,  Center for Computational Research /Buffalo,  TIGRE,  FermiGrid Grid Projects  DISUN  CDIGS National Grids: TeraGrid, HEP-Brazil International Grids: EGEE

6 R. Pordes, I Brazilian LHC Computing Workshop 6 What is an OSG Job? “ work done ” accomplished by and delivered as “ benefit received ” ; accountable to multiple organizations OSG EGEE Job Counted on Campus Grid, OSG and EGEE. MyApplication, EGEE RB,, VDS, OSG RESS Job Submission Condor-G Job does work benefiting WLCG.

7 R. Pordes, I Brazilian LHC Computing Workshop 7 Common Middleware provided through Virtual Data Toolkit Domain science requirements. OSG stakeholders and middleware developer (joint) projects. Integrate into VDT Release. Deploy on OSG integration grid Include in OSG release & deploy to OSG production. Globus, Condor, EGEE etc Test on “VO specific grid”

8 R. Pordes, I Brazilian LHC Computing Workshop 8 Reliable: Central Operations Activities Automated validation of basic services and site configuration Configuration of HeadNode and Storage to reduce errors:  Remove dependence on Shared File System  Condor-managed GRAM fork queue Scaling tests of WS-GRAM and GridFTP. Daily Grid Exerciser:

9 R. Pordes, I Brazilian LHC Computing Workshop 9 OSG Drivers:  Research groups transitioning from & extending (legacy) systems to Grids:  US LHC Collaborations  Contribute to & depend on milestones, functionality, capacity of OSG.  Commitment to general solutions, sharing resources & technologies;  Application Computer Scientists  Real life use of technology, integration, operation.  Federations with Campus Grids  Bridge & interface Local & Wide Area Grids.  Interoperation & partnerships with national/ international infrastructures  Ensure transparent and ubiquitous access.  Work towards standards. LIGO- gravitational wave physics; STAR - nuclear physics, CDF, D0, - high energy physics, SDSS - astrophysics GADU - bioinformatics Nanohub NMI, Condor, Globus, SRM GLOW, FermiGrid, GROW, Crimson, TIGRE EGEE, TeraGrid, INFNGrid

10 R. Pordes, I Brazilian LHC Computing Workshop 10 LHC Physics drive schedule and performance envelope  Beam starts in 2008:  Distributed System must serve 20PB of data in served across 30PB disk distributed across 100 sites worldwide to be analyzed by 100MSpecInt2000 of CPU.  Service Challenges give steps to full system 1 GigaByte/sec

11 R. Pordes, I Brazilian LHC Computing Workshop 11 Bridging Campus Grid Jobs - GLOW Dispatch jobs from local security, job, storage infrastructure and “uploading” to wide-area infrastructure. Fast ramp up in last week. Currently running the football pool problem which has application in data compression, coding theory, and statistical designs.

12 R. Pordes, I Brazilian LHC Computing Workshop 12 Genome Analysis and Database Update system Request: 1000 CPUs for 1-2 weeks. Once a month. 3 different applications: BLAST, Blocks, Chisel. Currently ramping up on OSG and receiving 600 CPUs and 17,000 jobs a week.

13 R. Pordes, I Brazilian LHC Computing Workshop 13 Common Middleware provided through Virtual Data Toolkit Domain science requirements. OSG stakeholders and middleware developer (joint) projects. Integrate into VDT Release. Deploy on OSG integration grid Include in OSG release & deploy to OSG production. Globus, Condor, EGEE etc Test on “VO specific grid” Condor project

14 R. Pordes, I Brazilian LHC Computing Workshop 14 of course a special grid … it’s the people… (some of them at the consortium meeting in Jan 06)

15 R. Pordes, I Brazilian LHC Computing Workshop 15 TeraGrid Through high-performance network connections, TeraGrid integrates high-performance computers, data resources and tools, and high-end experimental facilities around the (US) country. CDF MonteCarlo jobs running on Purdue TeraGrid resource; able to access OSG data areas and be accounted to both Grids. http://www.nsf.gov/news/news_images.jsp?cntn_id=104248&org=OLPA

16 R. Pordes, I Brazilian LHC Computing Workshop 16 OSG: More than a US Grid Taiwan - (CDF, LHC) Brazil - (D0, STAR, LHC) Korea

17 R. Pordes, I Brazilian LHC Computing Workshop 17 OSG: Where to find information: OSG Web site: www.opensciencegrid.orgwww.opensciencegrid.org Work in progress: http://osg.ivdgl.org/twiki/bin/view/Integration/Over viewGuide http://osg.ivdgl.org/twiki/bin/view/Integration/Over viewGuide Virtual Data Toolkit: http://vdt.cs.wisc.edu//index.html http://vdt.cs.wisc.edu//index.html News about Grids in Science in “Science Grid This Week”: www.interactions.org/sgtwwww.interactions.org/sgtw OSG Consortium meeting Seattle Aug 21st. Thank you!

18 R. Pordes, I Brazilian LHC Computing Workshop 18 BDII LDAP URLs OSG - EGEE Interoperation for WLCG Jobs GRAM T2 GRAM T2 GRAM T2 SRM GRAM T2 SRM Site SRM T2 GRAM T2 SRM GRAM T2 SRM T2 SRM Site SRM VO RB VO UI BDII Data Stores Picture thanks to I. Fisk

19 R. Pordes, I Brazilian LHC Computing Workshop 19 Open Science Grid in 1 minute: OSG Resources - use and policy under owner control. Clusters and storage shared across local, Campus intra-grid, Regional Grid and large federated Inter-Grids. OSG Software Stack - based on Virtual Data Toolkit. Interfaces:  Condor-G job submission interface;  GridFTP data movement  SRM storage management;  Glue Schema V1.2; easy to configure GIPs;, CEMON coming in 3 months. OSG Use - Register VO with with Operations Center; Provide  URL for VOMS service - this must be propagated to sites.  Contact for Support Center.  Join operations groups. OSG Job Brokering, Site Selection - no central or unique service.  LIGO uses Pegasus;  SDSS uses VDS;  STAR uses Star-schedule;  CMS uses EGEE-RB;  ATLAS uses Panda;  CDF uses CDF GlideCAF;  D0 uses SAM-JIM;  GLOW uses “condor-schedd on the side”.  Nano-hub uses application portal. OSG Storage & Space Management shared file systems; persistent VO application areas; SRM interfaces. OSG Operations - Distributed including each VO, Campus Grid. Operations is also a WLCG ROC. OSG Accounting & Monitoring - MonaLisa; can support rGMA; OSG meters/probes for Condor being released soon. US Tier-1s reporting monthly to WLCG APEL.

20 Services to the US Tier- 1 Sites LHCOPN April 4th, 2006 Joe Metzger metzger@es.net ESnet Engineering Group Lawrence Berkeley National Laboratory

21 R. Pordes, I Brazilian LHC Computing Workshop 21 ESnet Target Architecture: High-reliability IP Core Chicago Atlanta Seattle Albuquerque IP Core LA Denver Primary DOE Labs Possible hubs SDN hubs IP core hubs Washington DC Sunnyvale New York San Diego Cleveland

22 R. Pordes, I Brazilian LHC Computing Workshop 22 ESnet Target Architecture: Science Data Network New York Chicago Atlanta Seattle Albuquerque Science Data Network Core San Diego LA Sunnyvale Denver Primary DOE Labs Possible hubs SDN hubs IP core hubs Cleveland Washington DC

23 R. Pordes, I Brazilian LHC Computing Workshop 23 10-50 Gbps circuits Production IP core Science Data Network core Metropolitan Area Networks International connections Metropolitan Area Rings ESnet Target Architecture: IP Core+Science Data Network Core+Metro Area Rings New York Chicago Washington DC Atlanta Seattle Albuquerque San Diego LA Sunnyvale Denver Loop off Backbone SDN Core IP Core Primary DOE Labs Possible hubs SDN hubs IP core hubs international connections Cleveland


Download ppt "R. Pordes, I Brazilian LHC Computing Workshop 1 What is Open Science Grid?  High Throughput Distributed Facility  Shared opportunistic access to existing."

Similar presentations


Ads by Google