Download presentation
Presentation is loading. Please wait.
Published byEverett Simon Modified over 9 years ago
1
Prof. Srinidhi Varadarajan Director Center for High-End Computing Systems
2
We need a paradigm shift to make supercomputers more usable for mainstream computational scientists. A similar shift occurred in computing in the 1970s when the advent of inexpensive minicomputers into academia spurred a large body of computing research. Results from this research went back to industry creating a growth cycle that lead computing being a commodity. This requires a comprehensive “rethink” of programming languages, runtime systems, operating systems, scheduling, reliability and operations and management Moving to petascale and exascale class systems significantly complicates this challenge. Need a computing environment that can efficiently and usably span the scales from department sized systems to national resources.
3
The majority of our supercomputers today are distributed memory systems that use the message passing model of parallel computation. The shared/distributed memory view is a dichotomy imposed by hardware constraints. Modern high performance interconnects such as Infiniband are memory based systems. Provides the hardware basis to envision DSM systems that deepen the memory hierarchy. Most common operations are accelerated through hardware offload.
4
Common question. My application runs on my desktop, but it takes too long. Can I just run it on the supercomputer and make it run faster? Short answer: no. Longer answer: almost certainly not. As core frequencies have flattened, multi-core and many core architectures are here to stay. This is increasing the prevalence of threaded codes. Can we take standard threaded codes and run them on a cluster supercomputer without any modifications or recompilation?
5
The goal of our work is to enable Pthread based threaded codes to transparently run cluster supercomputers. The DSM system acts as the runtime, and provides a globally consistent memory abstraction. New consistency algorithm with release consistency semantics guarantees correct operation for valid threaded codes. No, it won’t fix your bugs, but it may make deadlock and livelock detection easier, possibly even automatic.
6
Separation of concerns The system is divided into a consistency layer and a lower level communication layer. The communication layer uses a well-defined architecture similar to MPI’s ADI to enable a wide variety of lower level interconnects. System consists of either dedicated memory servers or nodes may share a portion of their memory into a global pool. Dedicated memory servers are essentially low end servers that can host a lot of memory over a fast interconnect. Memory striping algorithms employed to mitigate memory access hotspots.
7
The DSM architecture uses a global scheduler that treats cluster nodes as a set of processors. Thread migration is simple and relatively inexpensive. This enables load balancing through runtime migration. Two issues: Compute Load Imbalance Data Affinity
8
Extending the threads model to support adaptivity. Transactional Memory An artifact of our consistency protocol enables us to provide transactional memory semantics fairly inexpensively. This enables speculative and/or adaptive execution models, particularly in hard to parallelize sequential sections of code. Speculation enables us to explore multiple execution paths, with with the DSM guaranteeing that there are no memory side effects. Invalid paths are simply pruned. Adaptive execution enables optimistic and conservative algorithms to be started concurrently.
9
Current threaded and message passing models are inadequate for peta and exascale systems. Growth in heterogeneous multi-core systems significantly complicates this problem. Need more comprehensive runtime systems that can aid in load balancing, profile guided optimization and code adaptation. The move in the compilers community towards greater emphasis on dynamic analysis is a step in this direction.
10
We are working on hybrid programming models that combine von Neumann program counter based elements embedded in dataflow constructs. A new model must provide insights into problem decomposition as well as map existing decomposition methods. Coordination models that can operate at peta and exascale.
11
Methods to evolve applications easily when requirements change. Working with the compilers, programming languages, architectures, software engineering, applications and systems communities to realize this goal.
12
System G: 2600 core Intel x86 Xeon Penryn processors with Quad Data Rate Infiniband. 12,000 thermal sensors, 5000 power sensors System X: 2200 processor PowerPC cluster with Infiniband interconnect. Anantham: 400 processor Opteron Cluster with Myrinet interconnect Several 8-32 processor research clusters. 12 processor SGI Altix shared memory system 8 processor AMD Opteron shared memory system. 16 core AMD Opteron shared memory system 16 node Playstation 3 cluster
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.