Presentation is loading. Please wait.

Presentation is loading. Please wait.

PRAGMA Institute, 9/26/2007 Cindy Zheng, David Abramson, Peter Arzberger, Shahaan Ayyub, Colin Enticott, Slavisa Garic, Wojtek Goscinski, Mason J. Katz,

Similar presentations

Presentation on theme: "PRAGMA Institute, 9/26/2007 Cindy Zheng, David Abramson, Peter Arzberger, Shahaan Ayyub, Colin Enticott, Slavisa Garic, Wojtek Goscinski, Mason J. Katz,"— Presentation transcript:

1 PRAGMA Institute, 9/26/2007 Cindy Zheng, David Abramson, Peter Arzberger, Shahaan Ayyub, Colin Enticott, Slavisa Garic, Wojtek Goscinski, Mason J. Katz, Bu Sung Lee, Phil M. Papadopoulos, Sugree Phatanapherom, Somsak Sriprayoonsakul, Yoshio Tanaka, Yusuke Tanimura, Osamu Tatebe, Putchong Uthayopas and the whole PRAGMA Grid team Pacific Rim Application and Grid Middleware Assembly PRAGMA Grid – Lessons Learned

2 PRAGMA Institute, 9/26/2007 Overview PRAGMA PRAGMA Grid –People –Hardware –Software –Operations Grid Applications Grid Middleware –Security –Infrastructure –Services Grid Interoperations Heterogeneity People Collaborations Integrations Lessons learned

3 PRAGMA Institute, 9/26/2007 pragma Overarching Goals PRAGMA A Practical Collaborative Framework People and applications Strengthen Existing and Establish New Collaborations Work with Science Teams to Advance Grid Technologies and Improve the Underlying Infrastructure In the Pacific Rim and Globally

4 PRAGMA Institute, 9/26/2007 PRAGMA Member Institutions KU NECTEC TNGC Thailand UoHyd India MIMOS USM Malaysia ASGCC NCHC Taiwan AIST CCS CMC NARC OsakaU TITech Japan BII IHPC NGO Singapore MU Australia APAC Australia JLU China CalIT2 CRBS SDSC UCSD USA CICESE Mexico NCSA StarLight TransPAC2 USA CNIC China CRAY PNWG USA KBSI KISTI Konkuk Korea APAN Japan 37 institutions from 12 countries/regions Founded 2002 Supported by Members UUtah USA BeSTGRID New Zealand

5 PRAGMA Institute, 9/26/2007 Overview and Approach Process to Promote Routine Use Team Science Application-Driven Collaborations ApplicationsMiddleware Routine Use Lab/Testbed Testing Applications Building Grid and GOC Multiway Dissemination Key Middleware Workshops and Organization Information Exchange Planning and Review New Collaborations New Members Expand Users Expand Impact Outcomes Improved middleware Broader Use New Collaborations Transfer Tech. Standards Publications New Knowledge Data Access Education

6 PRAGMA Institute, 9/26/2007 PRAGMA Working Groups Bioscience Telescience Geo-science Resources and data –Grid middleware interoperability –Global grid usability and productivity PRAGMA Grid effort is led by resources and data working group, but rely on collaborations and contributions among all working groups.

7 PRAGMA Institute, 9/26/2007 PRAGMA Grid 32 institutions in 16 countries/regions, 27 compute sites, 14 Gfarm sites (+ 6 in preparation) UZH Switzerland NECTEC ThaiGrid Thailand UoHyd India MIMOS USM Malaysia CUHK HongKong ASGC NCHC Taiwan HCMUT IOIT-HCM Vietnam AIST OsakaU UTsukuba TITech Japan BII IHPC NGO NTU Singapore MU Australia APAC QUT Australia KISTI Korea JLU China SDSC USA CICESE Mexico UNAM Mexico UCN Chile UChile Chile UUtah USA NCSA USA BU USA ITCR Costa Rica BESTGrid New Zealand CNIC GUCAS China LZU China UPRM Puerto Rico AIST SDSC NGO NECTEC ThaiGrid ASGC CNIC GUCAS MIMOS NCSA LZU IOIT-HCM CUHK UZH 14 gfarm sites

8 PRAGMA Institute, 9/26/2007 PRAGMA Grid Members and Team Sites –23 sites from PRAGMA member institutions –15 sites from Non-PRAGMA member institutions –27 sites contributed compute clusters Team members –160 and growing –one management contact / site –1~3 technical support contact / site –1~4 application drivers / application –1~5/Middleware development teams

9 PRAGMA Institute, 9/26/2007 PRAGMA Grid Compute Resources

10 PRAGMA Institute, 9/26/2007 Characteristics of PRAGMA Grid Grass-root –Voluntary contribution –Open (PRAGMA member or not, pacific rim or not) –Long-term collaborative working experiment Heterogeneous –Funding –No uniform infrastructure management –Variety of sciences and applications –Site policies, system and network environments Realistically tough –Good for development, collaborations, integrations and testing

11 PRAGMA Institute, 9/26/2007 PRAGMA Grid Software Layers Local job scheduler (require one) Globus (required) Application Middleware Applications Infrastructure Middleware SGEPBSLSFSQMS… Ninf-GNimrod/GMpich-GX…GfarmSCMSWebMOGASCSF Phylogenetic…FMOCSTFTSavannahMM5AMBERSiesta …

12 PRAGMA Institute, 9/26/2007 PRAGMA Grid Operations

13 PRAGMA Institute, 9/26/2007 One of the major lessons from PRAGMA Grid, that everybody has noticed and would agree– You have to Grid People before you can Grid machines Rajesh Chhabra Australia

14 PRAGMA Institute, 9/26/2007 Grid Operation Develop and maintain mutual beneficial and happy relationships among all people involved –Geographies, time-zones, languages –Funding, chain-of-command, priorities –Mutual benefit, consensus, active leadership –Coordinator, site contacts Collaboration tools –Mailing lists, VTCs, Skype, semi-annual workshops –Grid Operation Center (GOC) –Wiki, all sites and application, middleware teams collaborate Heterogeneity –Tolerate, technology, overcome and take advantage –Software inventory instead of software stack –Many sub-grids for applications –Recommendation instead of requirements –Software license

15 PRAGMA Institute, 9/26/2007 Create New Ways To Operate Lack precedence Everyone contributes ideas, suggestions Evolving and improving over time Everyone document and update (wiki) –Create new procedures New site setup to join PRAGMA Grid New user/application to run in PRAGMA grid doc/userguide/pragma_user_guide.html –Tabulate information Application pages, site pages, resources tables, status pages –Publish instructions Software deployment procedures, tools

16 PRAGMA Institute, 9/26/2007 Application Driven

17 PRAGMA Institute, 9/26/2007 Applications and Middleware Real science applications paired with and drive middleware development Open to applications of all scientific disciplines Achieve long-run and scientific results ~30 applications in 3 years: –Climate simulation Savannah/Nimrod (MU, Australia) MM5/Mpich-Gx (CICESE, Mexico; KISTI, Korea) –Quantum-mechanics, quantum-chemistry: TDDFT, QM-MD, FMO/Ninf-G (AIST, Japan) –Genomics and meta-genomics iGAP/Gfarm/CSF (UCSD, USA; AIST, Japan; JLU, China) HPM: genomics (IOIT-HCM, Vietnam) mpiBlast/Mpich-G2 (ASGC, Taiwan) Phylogenetic/Gfarm/CFS (UWisc and UCSD, USA) –Computational chemistry and fluid dynamics CSE-Online (UUtah, USA) e-AIRS (KISTI, Korea) Gamess-APBS/Nimrod (UZurich, Switzerland) –Molecular simulation Siesta/Nimrod (UZurich, Switzerland; MU, Australia) Amber/Gfarm ( USM, Malaysia; AIST, Japan) –Environmental Science CSTFT/Ninf-G (UPRM, Puerto Rico) –Computer Science Load Balancer (VAST-HCM, Vietnam) GriddLeS (MU, Australia)

18 PRAGMA Institute, 9/26/2007 Applications By PRIME Students Providing UCSD undergraduate students international interdisciplinary research internships and Cultural experiences since Sample applications ran in PRAGMA grid this year: –Climate modeling –Multi-walled carbon nanotube and polyethylene oxide composite computer visualization model –Metabolic regulation of ionic currents and pumps in rabbit ventricular myocyte model –Improving binding energy using quantum mechanics –Cardiac mechanics modeling –H5N1 simulation –Shp2 Protein Tyrosine Phosphatase Inhibitor simulation for cancer research

19 PRAGMA Institute, 9/26/2007 PRAGMA grid and its heterogeneous environment is great for –Testing –Collaborating –Integrating –Sharing Not easy –Middleware needs improvements Work in heterogeneous environment Fault tolerance –Need user friendly portals and services Automate and integrate –Information collections (grid monitoring, workflow) –Decisions and executions (scheduling) –Domain specific easy user interfaces (portals, CE tools) –… Lessons Learned From Running Applications

20 PRAGMA Institute, 9/26/2007 Grid Middleware

21 PRAGMA Institute, 9/26/2007 Ninf-G Developed by AIST, Japan Based on GridRPC model Support parallel computing OGF standard Integrated to NMI release 8 (first non-US software in NMI) Integrate with Rocks 4 applications ran in PRAGMA grid, 2 ran in multi-grid –TDDFT –QM/MD –FMO –CSTFT (UPRM) Achieved long runs (50 days) Improved fault-tolerance Simplified deployment procedures Speed-up development cycles

22 PRAGMA Institute, 9/26/2007 Nimrod/G Developed by Monash University (MU), Australia Supports large scale parameter sweeps on Grid infrastructure Easy user interface – Nimrod portals –MU, Australia –UZurich, Switzerland –UCSD, USA 3 applications ran in PRAGMA grid and 1 runs in multi-grids –Savanah climate simulation (MU) –GAMESS/APBS (UZurich) –Siesta (UZurich) Developed interface to Unicore Achieved long runs (90 different scenarios of 6 weeks each Improved fault-tolerance (innovate time_step) Enhancements in data and storage handling Job 1Job 2Job 3 Job 4Job 5Job 6 Job 7Job 8Job 9 Job 10Job 11Job 12 Job 13Job 14Job 15 Job 16Job 17Job 18 Description of Parameters PLAN FILE 1:30pm – Tutorial by David Abramson, Blair Bethwaite

23 PRAGMA Institute, 9/26/2007 Mpich-Gx Mpich-GX –Korea Institute of Science and Technology Information (KISTI), Korea –Based on Mpich-g2 –Grid-enabled MPI, support Private IP Fault tolerance MM5 and WRF –CICESE, Mexico –Medium scale atmospheric simulation model Experiment –KGrid –WRF work well with MPICH-GX –MM5 experienced scaling problems with MPICH-GX when use more than 24 processors in a cluster –Functionality of the private IP is usable –Performance of the private IP is reasonable

24 PRAGMA Institute, 9/26/2007 MM5-WRF/Mpich-GX Experiment pluto eolo Hurricane Marty Simulation Santana Winds Simulation output CICESE Ensenada USA México SDSC Private IP support Fault Tolerance support Mpich-GX KGrid 4pm tomorrow – Tutorial by Oh-kyoung Kwon

25 PRAGMA Institute, 9/26/2007 Science Technologies Collaborations Integrations

26 PRAGMA Institute, 9/26/2007 PRAGMA is a great model and needs to be emulated. Has helped weaken barriers between different research groups across different continents and allowed people to trust and collaborate rather than compete. Arun Agarwal UoHyd, India

27 PRAGMA Institute, 9/26/2007 Grid security –Naregi (Japan), APGrid, GAMA (SDSC, USA) Grid infrastructure –Monitoring - SCMSWeb (ThaiGrid, Thailand) –Accounting - MOGAS (NTU Singapore) –Metascheduling - Community Scheduler Framework (JLU, China) –Cyber-environment - CSE-Online (UUtah, USA) –Rocks and middleware (SDSC, USA; …) Ninf-G, SCE, Gfarm, Bio, K*Rocks, Condor, … Science, datagrid, sensor, network –Biosciences – Avian Flu, portal, … –Gfarm-fuse (AIST, Japan) –GEON data network –GLEON sensor network –OptIPuter High performance networked TDW Telescience Collaborations With Science and Technology Teams

28 PRAGMA Institute, 9/26/2007 Grid Security Trust in PRAGMA grid, doc/certificates.html doc/certificates.html –IGTF distribution –Non-IGTF distribution (trust all PRAGMA Grid sites) APGrid PMA –One of three IGTF founding PMAs –Many PRAGMA grid sites are members –PRAGMA CA Naregi-CA –AIST, UCSD, UChile, UoHyd, UPRM PRAGMA CA (experimental and production) –Based on Naregi-CA –Catch-all CA for PRAGMA –Production CA is IGTF compliant Myproxy and VOMS services –APAC Work with GAMA –Integrate with Naregi-CA (Naregi, UCSD) –Integration with VOMS (AIST) –Add servelet for account management (UChile) Lessons learned Leverage resources, setups and expertise Balance and consider both security and easy access and use Get more user communities involved with grid security

29 PRAGMA Institute, 9/26/2007 Gfarm Grid File System AIST, UTsukuba, Open source development at Grid file system that Federates storage of each site Meta-server keeps track of file copies and locations Can be mounted from cluster nodes and clients (GfarmFS-FUSE) Parallel I/O, near site copy for scalable performance Replication for fault tolerance Use GSI authentication Easy application deployment, file sharing

30 PRAGMA Institute, 9/26/2007 PRAGMA Gfarm Datagrid - Compute Cluster

31 PRAGMA Institute, 9/26/2007 Testing with applications Igap (Gfarm, Japan, UCSD, USA; JLU, China) –Huge number of small files –High meta-data access overhead –Meta-data cache server –Dramatic improvements (44sec -> 3.54sec) AMBER (USM, Malaysia; Gfarm, Japan) –Remote Gfarm meta-server –Meta-server is bottle-neck –File sharing permission, security –2.0 improved performance –Use as a shared storage only Version 1.4 works well in local or regional grid GeoGrid, Japan CLGrid, Chile Integration SCMSWeb (ThaiGrid, Thailand) Rocks (SDSC, USA; UZH, Switzerland) Develop and Test GfarmFS-FUSE in PRAGMA Grid

32 PRAGMA Institute, 9/26/2007 SCMSWeb Developed by Kasetsart University and ThaiGrid Web-based real-time grid monitoring system –System usage, Job/queue status –Probe – Globus authentication, job submission, gridftp, Gfarm access, … –Network bandwidth measurements with Iperf –PRAGMA grid geo map Support Linux, Solaris. Good meta-view, easy user interface, excellent user support Develop and test in PRAGMA grid –Deployed in 27 sites, improve scalability and performance –Sites help with porting to ia64 and Solaris –Demands push fast expansion of functionalities More regional/national grids learned and adopted

33 PRAGMA Institute, 9/26/2007 SCMSWeb Collaborations and Integrations Grid Interoperation Now (GIN, OGF) –Worked with PRAGMA grid, TeraGrid, OSG, NorduGrid and EGEE on GIN testbed monitoring added probes to handle various grid service configurations/tests. –Worked with CERN and Implemented a XML-> LDIF translator for GIN geo map –Worked with many grid monitor software developers on a common schema for cross-grid monitoring Software integration and interoperations –Rocks – SCE roll –MOGAS, grid accounting –CSE-Online, CSF, provide resource info Things are being worked on and planned –Data federator for grid applications Provide site software information Standardize data extractions and formats Improve data storage with RDBMS –Interoperate with other monitoring software Ganglia support

34 PRAGMA Institute, 9/26/2007 MOGAS Multi-Organization Grid Accounting System (MOGAS) –Lead by NanYang University, funded by National Grid Office in Singapore –Build on globus core (gridftp, GRAM, GSI) –Support GT2,3,4, SGE, PBS –Job/user/cluster/OU/grid levels usages; job logs; metering and charging tools Develop and test in PRAGMA grid –Deployed on 14 sites: different GT versions, job schedulers, GRAM scripts, security policies –Feedbacks, improve, automate deployment procedure –Decentralized servers and better database to improve scalability and performance –Collaborations and integrations with applications and other middleware teams push the development of easy database interface 4pm – MOGAS tutorial by Francis Lee

35 PRAGMA Institute, 9/26/2007 Community Scheduler Framework, v4 – meta-scheduler –Developed by Jilin University, China –Grid services host in GT4, WSRF compliant, execution Component in Globus Toolkit 4 –Open Source, –Support GT2&4, LSF, PBS, SGE, Condor –Easy user interface - portal Testing and collaborating in PRAGMA –Testing with application iGAP (UCSD, AIST, KISTI, …) –Collaborate and integrate with Gfarm on data staging (AIST, Japan) –Setup a CSF server and portal (SDSC, USA) –Collaborate/integrate with SCMSWeb for resource information (Thaigrid, Thailand) –Leverage resources and global grid testing environment CSF4 1:30pm – CSF4 Tutorial by Zhao-hui Ding

36 PRAGMA Institute, 9/26/2007 Computational Science & Engineering Online Developed by University of Utah, USA (Thanh N. Truong) Desktop tool, user friendly interface enables seamless access to remote data, tools and grid computing resources Currently support computational Chemistry Can be customized for other domain science Developed interface to TeraGrid Collaborate with ThaiGrid as case study –Used for Computational workshop –Extend grid access to portal architecture –Improved security Working on interface PRAGMA grid –Heterogeneity Quantum ChemistryNano-materialsDrug Design

37 PRAGMA Institute, 9/26/2007 Collaborations with OptIPuter OptIPuter (Optical networking, Internet Protocol, computer storage, processing and visualization technologies) –Infrastructure that will tightly couple computational resources over parallel optical networks using the IP communication mechanism –central architectural element is optical networking, not computers –enable scientists who are generating terabytes and petabytes of data to interactively visualize, analyze, and correlate their data from multiple storage sites connected to optical networks Rocks VIS-roll (SDSC) Networked Tile Display Walls (TDW) –Low cost –For research collaboration –For remote education and conferencing –Deployed in PRAGMA grid 9 sites and more to follow Future plan –Global Lambda Integrated Facility (GLIF) –Solve grid application bandwidth problem CNIC, China UCSD, USA

38 PRAGMA Institute, 9/26/2007 OGF – GIN-OPS GIN testbed (February, 2006 – on-going) Application driven –TDDFT/Ninf-G (PRAGMA - AIST, Japan) PRAGMA, TeraGrid, OSG, NorduGrid; EGEE –Savanah fire simulation (PRAGMA – Monash University, Australia) PRAGMA, TeraGrid, OSG Multi-Grid monitoring –SCMSWeb probe matrix (PRAGMA - ThaiGrid, Thailand) –Common schema (PRAGMA, TeraGrid, EGEE, NorduGrid) Grid Interoperation Now (GIN)

39 PRAGMA Institute, 9/26/2007 OSG-PRAGMA Grid Interoperation Experiments More resources and support from each grid, but no special arrangements Application long-run –GridFMO/Ninf-G – Large scale quantum Chemistry (Tsutomo Ikegami, AIST, Japan) –240 CPUs from OSG and PRAGMA grid, 10 days x 7 calculations –Fault-tolerance enabled long-run –Meaningful and usable scientific results

40 PRAGMA Institute, 9/26/2007 Lessons Learned From Grid Interoperation Grid interoperation make large scale calculations possible Differences among grids provide learning, collaboration and integration opportunities –IGTF, VOMS (GIN) –Common Software Area (TeraGrid) –Ninf-G (AIST/PRAGMA) interface to NorduGrid –Nimrod-G (MU/PRAGMA) interface to Unicore (PRIME) –VDT (OSG) and Rocks (SDSC/PRAGMA) integration Differences in grid environment are source of difficulties for users and applications –Different user access setup procedure - take extra effort –Different job submission protocols GRAM, Sandbox, gridftp, modified GRAM, … One-to-one interface building is not scalable, nor desirable. Need standard. Middleware fault tolerance and flexible resource management is important

41 PRAGMA Institute, 9/26/2007 Collaborate in Publishing Research Results Some published papers in 2007: Amaro, RE, Minh DDL, Cheng LS, Lindstrom, WM Jr, Olson AJ, Lin JH, Li WW, and McCammon JA. Remarkable Loop Flexibility in Avian Influenza N1 and Its Implications for Antiviral Drug Design. J. AM. CHEM. SOC. 2007, 129, (PRIME) Choi Y, Jung S, Kim D, Lee J, Jeong K, Lim SB, Heo D, Hwang S, and Byeon OH."Glyco-MGrid: A Collaborative Molecular Simulation Grid for e-Glycomics," in 3rd IEEE International Conference on e-Science and Grid Computing, Banglore, India, Accepted. Ding Z, Wei W, Luo Y, Ma D, Arzberger PW, and Li WW, "Customized Plug-in Modules in Metascheduler CSF4 for Life Sciences Applications," New Generation Computing, p. In Press, Ding Z, Wei S, Ma, D and Li WW, "VJM -- A Deadlock Free Resource Co-allocation Model for Cross Domain Parallel Jobs," in HPC Asia 2007, Seoul, Korea, 2007, p. In Press. Görgen K, Lynch H, Abramson D, Beringer J and Uotila P. "Savanna fires increase monsoon rainfall as simulated using a distributed computing environment", to appear, Geophysical Research Letters. Ichikawa K, Date S, Krishnan S, Li W, Nakata K, Yonezawa Y, Nakamura H, and Shimojo S, "Opal OP: An extensible Grid-enabling wrapping approach for legacy applications", GCA Proceedings of the 3rd workshop on Grid Computing & Applications -, pp , Singapore, June 2007 a. (PRIUS) Ichikawa K, Date S, and Shimojo S. A Framework for Meta-Scheduling WSRF Based Services, Proceedings of 2007 IEEE Pacific Rim Conference on Communications, Computers and Signal Processing (PACRIM 2007), Victoria, Canada, pp , Aug b. (PRIUS) Kuwabara S, Ichikawa K, Date S, and Shimojo S. A Built-in Application Control Module for SAGE, Proceedings of 2007 IEEE Pacific Rim Conference on Communications, Computers and Signal Processing (PACRIM 2007), Victoria, Canada, pp , Aug (PRIUS) Takeda S, Date S, Zhang J, Lee BU, and Shimojo S. Security Monitoring Extension For MOGAS, GCA Proceedings of the 3rd workshop on Grid Computing & Applications -, pp Singapore, June (PRIUS) Tilak S, Hubbard P, Miller M, and Fountain T, ``The Ring Buffer Network Bus (RBNB) DataTurbine Streaming Data Middleware for Environmental Observing Systems," to appear in the Proceedings of the e-Science 2007 Zheng C, Katz M, Papadopoulos P, Abramson D, Ayyub S, Enticott C, Garic S, Goscinski W, Arzberger P, Lee B S, Phatanapherom S, Sriprayoonsakul S, Uthayopas P, Tanaka Y, Tanimura Y, Tatebe O. Lesson Learned Through Driving Science Applications in the PRAGMA Grid. Int. J. Web and Grid Servies, Vol.3, No.3, pp …

42 PRAGMA Institute, 9/26/2007 Summary PRAGMA grid –Shared vision lower resistance to use others software, test on others resources –Formed new development collaborations –Size and heterogeneity, explore issues which functional grid must resolve Management, resources and software coordination Identity and fault management Scalability and performance Feedback between application and middleware help improve software and promote software integration Heterogeneous global grid –Is realistic and challenge –Can be good for middleware development and testing –Can be useful for real science Impact –Software dissemination (Rocks, Ninf-G, Nimrod, SCMSWeb, Naregi-CA, …) –Help new national/regional grids (Chile, Vietnam, Hong kong, …) Key is people, is collaboration

43 PRAGMA Institute, 9/26/2007 A Grass Roots Effort One of the most important lessons of the Internet is that it grows most successfully where grass roots initiatives are encouraged and enabled. The Internet has historically grown from the bottom up, and this aspect continues to fuel its continued growth in the academic and commercial sectors. –Vint Cert, UN Economic and Social Council in 2000

44 PRAGMA Institute, 9/26/2007 pragma Thank You PRAGMA is supported by the National Science Foundation (Grant No. INT , INT , OCI ) and by member institutions PRIME is supported by the National Science Foundation under NSF INT PRAGMA grid is the result of contributions and support from all PRAGMA grid team members

Download ppt "PRAGMA Institute, 9/26/2007 Cindy Zheng, David Abramson, Peter Arzberger, Shahaan Ayyub, Colin Enticott, Slavisa Garic, Wojtek Goscinski, Mason J. Katz,"

Similar presentations

Ads by Google