Presentation is loading. Please wait.

Presentation is loading. Please wait.

NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth.

Similar presentations


Presentation on theme: "NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth."— Presentation transcript:

1

2 NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth Structure PI: T. H. Jordan (USC); Co-PIs: Y. Cui (SDSC), K. Olsen (SDSU), and J. Tromp (Princeton) Year- 1 plan is a set-up and demonstration phase comprising seven principal tasks: Task 1.1. Assemble community computational platforms from existing software components and deploy them at NWSC. Task 1.2. Optimize computational performance of AWP-ODC and SPECFEM3D codes on NWSC Yellowstone supercomputer. Task 1.3. Adapt Pegasus-WMS to support file management on community computational platforms. Task 1.4. Synthesize existing California CVM components and publish a statewide starting CVM for full- 3D inversion Task 1.5. Cross-validate the AWP-ODC and SPECFEM3D platforms. Task 1.6. Preserve constraints on CVM shallow structure during tomographic inversions. Task 1.7. Demonstrate capabilities for adjoint tomography on a global scale. The Year-2 plan is a production and delivery phase comprising seven principal tasks: Task 2.1. Exploit heterogenous petascale architectures for GPU-based accelerations of AWP-ODC and SPECFEM3D codes and verify performance by executing standard forward problems Task 2.2. Establish automated scientific workflows for full-3D inversions on the AWP-ODC and SPECFEM3D platforms. Task 2.3. Produce statewide California CVMs by full-3D inversions of earthquake, ambient-noise, and prior-constraint data on the AWP-ODC and SPECFEM3D platforms. Task 2.4. Validate full-3D tomography through UCVM-based comparisons of California inversion results from the AWP-ODC and SPECFEM3D platforms. Task 2.5. Deploy federated data management tools at NWSC and SCEC data centers for managing life- cycle of community data collections. Task 2.6. Complete the first phase of global adjoint tomography. Task 2.7. Publish improved statewide California CVMs for use in CyberShake hazard modeling. 2

3

4

5

6 Figure 3: (a) Map of topography and major faults (thick black lines) of southern California. (b) The optimal perturbation results of the southern California tomographic inversion including iteration CVM-S4.21 performed on Yellowstone. In perturbation maps, the red regions represent velocity reduction areas and the blue regions represent velocity increase areas. 6

7 ANGF examples cross southern Great Valley

8 8 CVM4 VS 20km 3.8 km/s ± 10% Perturbation 10km VS of CVM4 ± 15% CVM4 VS 10km 3.6 km/s ± 15% CVM4SI22 VS 10km 3.6 km/s ± 15% CVM4SI22 VS 20km 3.8 km/s ± 10% Perturbation 20km VS of CVM4 ± 10%

9 9 CVM4 CVM4SI22 NC

10 Perform two sets of 150 simulations for the fine and coarse mesh. Using the two sets of synthetics, thoroughly document the resolvable periods. This will dictate what bandpass will be used for measurements in the inversion. Move forward with CVM-H inversion, with emphasis on the uppermost 10 km and at a numerical resolution of 2 s. Model enhancements – 3D adjoint waveform tomography Tape et al., 2013 Seismogram-based estimates of the resolvable period

11 11 Probabilistic Seismic Hazard Analysis What will peak ground motion be over the next 50 years? –Used in building codes, insurance, government, planning –Answered via Probabilistic Seismic Hazard Analysis (PSHA) –Communicated with hazard curves and maps Hazard curve for downtown LA 2% in 50 years 0.6 g Probability of exceeding 0.1g in 50 yrs

12 CyberShake Study 13.4 Interested in velocity model, SGT code contribution to PSHA Planned CyberShake run –286 locations in Southern California –4 permutations of velocity model, SGT code –Use Blue Waters, Stampede, HPCC

13 13 CyberShake workflows Tensor extraction Seismogram synthesis Tensor extraction Tensor simulation...... x7,000 x415,000 x1 Seismogram synthesis Mesh generation Tensor Workflow x1x2 Post-Processing Workflow...... Hazard Curve

14 14 Scientific Workflows Large-scale, heterogeneous, high throughput –Parallel and many (~415,000) serial tasks Automation Data management Error recovery Resource provisioning Scalable We use Pegasus-WMS, HTCondor, Globus

15 Workflow Tool Development to Support CyberShake We started with excellent scientific codes and improved it over 5 years. In the following section, we describe a number of late-stage improvements that enabled us to reach the M8 milestone. 15

16 Pegasus-mpi-cluster Ships with Pegasus-WMS MPI wrapper around serial or thread-parallel jobs –Master-worker paradigm –Preserves dependencies –Specify jobs as usual, Pegasus does wrapping Uses intelligent scheduling –Core counts, memory requirements, priorities –Locality preferences under development Can aggregate output –Master collects worker output, writes in large chunks 16

17 17 Computational Requirements ComponentDataExecutionsCores/execCPU hours Mesh generation15 GB116050 Tensor simulation40 GB24,0008,000 Tensor extraction690 GB7,0001200 Seismogram synthesis 10 GB415,0001800 Curve generation1 MB11< 1 Total755 GB422,0009,000 Tensor Creation Post Processing This is for one location of interest; want to run >1000

18 CyberShake Study 13.4 Performance April 17, 2013 – June 17, 2013 Blue Waters (MPI SGT workflows): –Average of 19,300 cores, 8 jobs Stampede (HTC post-processing workflows): –Average of 1,860 cores, 4 jobs –470 million tasks executed (177 tasks/sec) –21,912 jobs total Managed 830 TB of data –57 TB output files –12.3 TB staged back to HPCC (~16M files) –1.5 billion rows added to database

19 19 CyberShake Study 13.4 Results Ratio comparison of SGT codes Ratio comparison of velocity models

20 Future Directions Migrate to GPU version of SGT code Create similar maps for third velocity model Increase frequency of calculations from 0.5 to 1 Hz – 16x for SGT calculations – 50x for post-processing Move to newest earthquake rupture forecast, UCERF 3.0 – 25x earthquakes to consider

21 For More Information Please Visit: http://scec.usc.edu/scecpedia End


Download ppt "NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth."

Similar presentations


Ads by Google