Presentation is loading. Please wait.

Presentation is loading. Please wait.

Please do not redistribute outside of the RACM group ① RACM2 and ARSC ② Nuts and bolts for RACM Presentation to RACM Meeting May 18, 2010 Please do not.

Similar presentations


Presentation on theme: "Please do not redistribute outside of the RACM group ① RACM2 and ARSC ② Nuts and bolts for RACM Presentation to RACM Meeting May 18, 2010 Please do not."— Presentation transcript:

1 Please do not redistribute outside of the RACM group ① RACM2 and ARSC ② Nuts and bolts for RACM Presentation to RACM Meeting May 18, 2010 Please do not redistribute outside of the RACM group

2 ① RACM2 and ARSC

3 Please do not redistribute outside of the RACM group Preamble ARSC’s involvement with RACM has been satisfying and worthwhileARSC’s involvement with RACM has been satisfying and worthwhile –Scientific progress –Community model building –Engagement of postdoctoral fellows –Engagement with multi-agency partners (DoD, DoE, TeraGrid and others) –Arctic theme relevant to ARSC & UAF ARSC would be pleased to be involved in the upcoming RACM2ARSC would be pleased to be involved in the upcoming RACM2

4 Please do not redistribute outside of the RACM group RACM2 Constraints While ARSC’s HPCMP (DoD) funding remains stable, the portion allocatable for science, versus HPCMP program support, is much more limitedWhile ARSC’s HPCMP (DoD) funding remains stable, the portion allocatable for science, versus HPCMP program support, is much more limited With the departure of postdocs He & Roberts, the “personnel” tie to RACM2 is not as clearWith the departure of postdocs He & Roberts, the “personnel” tie to RACM2 is not as clear

5 Please do not redistribute outside of the RACM group RACM2 “fit” ARSC, along with IARC and others at UAF, has enduring interests in many RACM/RACM2 themes:ARSC, along with IARC and others at UAF, has enduring interests in many RACM/RACM2 themes: –Climate modeling and a general high latitudes focus –Model scaling, coupling, and other practical issues –Enhancing aspects of land surface, ocean/ice and atmosphere models for arctic use

6 Please do not redistribute outside of the RACM group ARSC Benefits to RACM Large allocations of CPU hours on large systemsLarge allocations of CPU hours on large systems Capable & responsive user supportCapable & responsive user support Support of 2 postdoctoral fellows (Roberts & He)Support of 2 postdoctoral fellows (Roberts & He) Direct ties to expertise at UAF (IARC & elsewhere)Direct ties to expertise at UAF (IARC & elsewhere) Deep shared interests in RACM themesDeep shared interests in RACM themes

7 Please do not redistribute outside of the RACM group Potential ways Forward How can RACM2 best continue to engage with and benefit from ARSC? One or more of:How can RACM2 best continue to engage with and benefit from ARSC? One or more of: –PI/Co-I at UAF (ARSC or IARC or elsewhere) –RACM2 personnel budget to ARSC: consultant, specialist, staff scientist –RACM2 funding postdoc(s) at ARSC/IARC –RACM2 payment for cycles, storage. Approximately cents per CPU hour ( ), $1600/TB storage (including 2 nd copy) –What else???

8 Please do not redistribute outside of the RACM group Current Requirements for ARSC HPC Allocations See uire.html#AcademicSee uire.html#Academic ARSC project PI at UAFARSC project PI at UAF Tie to arctic researchTie to arctic research Note that the largest allocations (including “arscasm”) do get extra scrutiny and attention. Allocations over around 1M hours are in strategic areas, and in deep enduring interest areasNote that the largest allocations (including “arscasm”) do get extra scrutiny and attention. Allocations over around 1M hours are in strategic areas, and in deep enduring interest areas

9 Please do not redistribute outside of the RACM group Future Requirements for ARSC HPC Allocations ARSC will be adding UAF user community members to the allocations application review processARSC will be adding UAF user community members to the allocations application review process The Configuration Change Board of ARSC will, similarly, have user representationThe Configuration Change Board of ARSC will, similarly, have user representation UAF will, we think, provide budget to support ARSC. This might come with some expectations about access to systems that are at least partially UAF-supportedUAF will, we think, provide budget to support ARSC. This might come with some expectations about access to systems that are at least partially UAF-supported Newby will, we hope, get NSF funding that will allow growth of the academic HPC resources. This will include direct responsibility in what some of the ARSC resources will be available forNewby will, we hope, get NSF funding that will allow growth of the academic HPC resources. This will include direct responsibility in what some of the ARSC resources will be available for

10 Please do not redistribute outside of the RACM group ARSC HPC Systems: Now Midnight: core Linux cluster, SLES 9.3. HPCMP allocated, with some academic useMidnight: core Linux cluster, SLES 9.3. HPCMP allocated, with some academic use Pingo: 3456 core XT5. HPCMP allocated, with some academic usePingo: 3456 core XT5. HPCMP allocated, with some academic use Pacman: 528 core Linux cluster in pioneer testing. Academic only.Pacman: 528 core Linux cluster in pioneer testing. Academic only. Storage: Seawolf. HPCMP, with some academic use.Storage: Seawolf. HPCMP, with some academic use. Storage: Bigdipper. Academic only.Storage: Bigdipper. Academic only.

11 Please do not redistribute outside of the RACM group ARSC HPC Systems: October 1 Chugach: Cray XE6. HPCMP only (no academic use)Chugach: Cray XE6. HPCMP only (no academic use) Phoenix (code name): Reborn and smaller midnight, likely ~250 x2200m2 nodes, Linux clusterPhoenix (code name): Reborn and smaller midnight, likely ~250 x2200m2 nodes, Linux cluster Pacman: Potentially grown to ~1500 cores if Newby’s grant comes throughPacman: Potentially grown to ~1500 cores if Newby’s grant comes through February 2011: Grown further by year 2 EPSCoR/PACMAN grantFebruary 2011: Grown further by year 2 EPSCoR/PACMAN grant Storage: HPCMP storage unavailable for academic useStorage: HPCMP storage unavailable for academic use Storage: Bigdipper, with growth plans for disk and tapeStorage: Bigdipper, with growth plans for disk and tape Storage: If Newby’s grant comes through, additional disk for data-intensive computingStorage: If Newby’s grant comes through, additional disk for data-intensive computing –Worst case: phoenix plus 528-core pacman. This will still provide enough CPU hours for all of the current ARSC academic research at 2010 levels

12 Please do not redistribute outside of the RACM group ② Nuts and bolts for RACM

13 Please do not redistribute outside of the RACM group Academic Systems Considerations – Job Size Initial systems on academic side will be smaller than pingo.Initial systems on academic side will be smaller than pingo. –Maximum job size will likely be ~256 cores on pacman. –“Phoenix” (midnight reborn) maximum job size not yet determined, but unlikely to be more than 512 cores. –Additional funding for pacman system will increase core count later this year or early next year

14 Please do not redistribute outside of the RACM group $ARCHIVE data By October there will be distinct HPCMP and Academic systems being operated by ARSC.By October there will be distinct HPCMP and Academic systems being operated by ARSC. –Team members of both Academic (e.g. ARSCASM) and HPCMP (e.g. NPSCA242) need to do some planning to make sure data is accessible to others on the team. The ARSC Help Desk has sent messages to people in this category to ask them to set UNIX groups on $ARCHIVE data. –Need to be aware of data locality for pre/post processing. HPCMP systems will be in Illinois and Academic systems in Fairbanks, AK.

15 Please do not redistribute outside of the RACM group $ARCHIVE data continued Academic project $ARCHIVE data will be migrated to bigdipper over the summer.Academic project $ARCHIVE data will be migrated to bigdipper over the summer. –Due to DoD requirements we will not have $ARCHIVE on bigdipper available via NFS on midnight or pingo. –We may allow passwordless access from midnight and pingo. –bigdipper $ARCHIVE is available via NFS on pacman now and will be on other academic systems later this year.

16 Please do not redistribute outside of the RACM group Academic System Considerations Continued $WORKDIR on “Phoenix” and pacman are smaller than pingo.$WORKDIR on “Phoenix” and pacman are smaller than pingo. –Can’t support multi-TBs of use by project members on a continual basis. $ARCHIVE cache for bigdipper (new storage server) is much larger than seawolf.$ARCHIVE cache for bigdipper (new storage server) is much larger than seawolf. –Files should stay online longer. –Should have better NFS connectivity to $ARCHIVE. –May be able to do some pre/post processing right in $ARCHIVE.

17 Please do not redistribute outside of the RACM group Other Comments Help Desk support will stay the same as it has been previously.Help Desk support will stay the same as it has been previously. Software stack for HPC systems will likely be similar to what we have had with PGI compiler suite as the default. We will probably drop support for PathScale.Software stack for HPC systems will likely be similar to what we have had with PGI compiler suite as the default. We will probably drop support for PathScale.

18 Please do not redistribute outside of the RACM group Current RACM Usage Midnight October 1, 2009 through 15 May 2010.Midnight October 1, 2009 through 15 May Foreground Background Total Remaining ( 13.34%) Pingo October 1, 2009 through 15 May 2010.Pingo October 1, 2009 through 15 May Foreground Background Total Remaining ( 51.44%)


Download ppt "Please do not redistribute outside of the RACM group ① RACM2 and ARSC ② Nuts and bolts for RACM Presentation to RACM Meeting May 18, 2010 Please do not."

Similar presentations


Ads by Google