Parallel Computing Sciences Department MOV’01 Multilevel Combinatorial Methods in Scientific Computing Bruce Hendrickson Sandia National Laboratories Parallel.

Slides:



Advertisements
Similar presentations
05/11/2005 Carnegie Mellon School of Computer Science Aladdin Lamps 05 Combinatorial and algebraic tools for multigrid Yiannis Koutis Computer Science.
Advertisements

Multilevel Hypergraph Partitioning Daniel Salce Matthew Zobel.
Minimum Clique Partition Problem with Constrained Weight for Interval Graphs Jianping Li Department of Mathematics Yunnan University Jointed by M.X. Chen.
Fast Algorithms For Hierarchical Range Histogram Constructions
The Combinatorial Multigrid Solver Yiannis Koutis, Gary Miller Carnegie Mellon University TexPoint fonts used in EMF. Read the TexPoint manual before you.
Modularity and community structure in networks
CS 290H 7 November Introduction to multigrid methods
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Lecture 17 Introduction to Eigenvalue Problems
Lecture 21: Spectral Clustering
Agenda Seam-carving: finish up “Mid-term review” (a look back) Main topic: Feature detection.
1cs542g-term Sparse matrix data structure  Typically either Compressed Sparse Row (CSR) or Compressed Sparse Column (CSC) Informally “ia-ja” format.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Numerical geometry of non-rigid shapes
Implicit Hitting Set Problems Richard M. Karp Harvard University August 29, 2011.
Avoiding Communication in Sparse Iterative Solvers Erin Carson Nick Knight CS294, Fall 2011.
Segmentation Graph-Theoretic Clustering.
Chapter 9 Graph algorithms Lec 21 Dec 1, Sample Graph Problems Path problems. Connectedness problems. Spanning tree problems.
High Performance Computing 1 Parallelization Strategies and Load Balancing Some material borrowed from lectures of J. Demmel, UC Berkeley.
A scalable multilevel algorithm for community structure detection
Randomness in Computation and Communication Part 1: Randomized algorithms Lap Chi Lau CSE CUHK.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
Application of Graph Theory to OO Software Engineering Alexander Chatzigeorgiou, Nikolaos Tsantalis, George Stephanides Department of Applied Informatics.
15-853Page :Algorithms in the Real World Separators – Introduction – Applications.
Multilevel Graph Partitioning and Fiduccia-Mattheyses
Joanna Ellis-Monaghan, St. Michaels College Paul Gutwin, Principal Technical Account Manager, Cadence.
Multilevel Hypergraph Partitioning G. Karypis, R. Aggarwal, V. Kumar, and S. Shekhar Computer Science Department, U of MN Applications in VLSI Domain.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Gene expression & Clustering (Chapter 10)
Combinatorial Scientific Computing is concerned with the development, analysis and utilization of discrete algorithms in scientific and engineering applications.
Sandia National Laboratories Graph Partitioning Workshop Oct. 15, Load Balancing Myths, Fictions & Legends Bruce Hendrickson Sandia National Laboratories.
Graph Partitioning Donald Nguyen October 24, 2011.
Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy’s National Nuclear.
1 SOC Test Architecture Optimization for Signal Integrity Faults on Core-External Interconnects Qiang Xu and Yubin Zhang Krishnendu Chakrabarty The Chinese.
Segmentation Course web page: vision.cis.udel.edu/~cv May 7, 2003  Lecture 31.
CSCE350 Algorithms and Data Structure Lecture 17 Jianjun Hu Department of Computer Science and Engineering University of South Carolina
Graph Algorithms. Definitions and Representation An undirected graph G is a pair (V,E), where V is a finite set of points called vertices and E is a finite.
CS 290H Lecture 5 Elimination trees Read GLN section 6.6 (next time I’ll assign 6.5 and 6.7) Homework 1 due Thursday 14 Oct by 3pm turnin file1.
Combinatorial Scientific Computing and Petascale Simulation (CSCAPES) A SciDAC Institute Funded by DOE’s Office of Science Investigators Alex Pothen, Florin.
NP-COMPLETE PROBLEMS. Admin  Two more assignments…  No office hours on tomorrow.
Spectral Sequencing Based on Graph Distance Rong Liu, Hao Zhang, Oliver van Kaick {lrong, haoz, cs.sfu.ca {lrong, haoz, cs.sfu.ca.
NP-Complete problems.
Data Structures and Algorithms in Parallel Computing Lecture 2.
PaGrid: A Mesh Partitioner for Computational Grids Virendra C. Bhavsar Professor and Dean Faculty of Computer Science UNB, Fredericton This.
CS 484 Load Balancing. Goal: All processors working all the time Efficiency of 1 Distribute the load (work) to meet the goal Two types of load balancing.
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
Data Structures and Algorithms in Parallel Computing Lecture 7.
CS 290H Administrivia: May 14, 2008 Course project progress reports due next Wed 21 May. Reading in Saad (second edition): Sections
Graph Partitioning using Single Commodity Flows
Chapter 13 Backtracking Introduction The 3-coloring problem
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Multilevel Partitioning
Spectral Clustering Shannon Quinn (with thanks to William Cohen of Carnegie Mellon University, and J. Leskovec, A. Rajaraman, and J. Ullman of Stanford.
CSCAPES Mission Research and development Provide load balancing and parallelization toolkits for petascale computation Develop advanced automatic differentiation.
CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning
High Performance Computing Seminar II Parallel mesh partitioning with ParMETIS Parallel iterative solvers with Hypre M.Sc. Caroline Mendonça Costa.
High Performance Computing Seminar
Parallel Hypergraph Partitioning for Scientific Computing
Solving Linear Systems Ax=b
A Continuous Optimization Approach to the Minimum Bisection Problem
CS 290H Administrivia: April 16, 2008
Segmentation Graph-Theoretic Clustering.
Randomized Algorithms CS648
Carlos Ordonez, Predrag T. Tosic
3.3 Network-Centric Community Detection
Read GLN sections 6.1 through 6.4.
Presentation transcript:

Parallel Computing Sciences Department MOV’01 Multilevel Combinatorial Methods in Scientific Computing Bruce Hendrickson Sandia National Laboratories Parallel Computing Sciences Dept.

Parallel Computing Sciences Department MOV’01 An Overdue Acknowledgement l Parallel Computing Uses Graph Partitioning l We owe a deep debt to circuit researchers »KL/FM »Spectral partitioning »Hypergraph models »Terminal propagation

Parallel Computing Sciences Department MOV’01 In Return … l We’ve given you »Multilevel partitioning »hMETIS l Our applications are different from yours »Underlying geometry »More regular structure »Bounded degree »Partitioning time is more important –Different algorithmic tradeoffs

Parallel Computing Sciences Department MOV’01 Multilevel Discrete Algorithm l Explicitly mimic traditional multigrid l Construct series of smaller approximations »Restriction l Solve on smallest »Coarse grid solve l Propagate solution up the levels »Prolongation l Periodically perform local improvement »Smoothing

Parallel Computing Sciences Department MOV’01 Lots of Possible Variations l More complex multilevel iterations »E.g. V-cycle, W-cycle, etc. »Not much evidence of value for discrete problems l Key issue: properties of coarse problems »Local refinement = multi-scale improvement l I’ll focus on graph algorithms »Most relevant to VLSI problems

Parallel Computing Sciences Department MOV’01 Not a New Idea l Idea is very natural »Reinvented repeatedly in different settings l Focus of this workshop is on heuristics for hard problems l Technique also good for poly-time problems »E.g. Geometric point detection (Kirkpatrick’83)

Parallel Computing Sciences Department MOV’01 Planar Point Detection l O(n log n) time to preprocess l O(log n) time to answer query

Parallel Computing Sciences Department MOV’01 Multilevel Graph Partitioning l Invented Independently Several Times »Cong/Smith’93 »Bui/Jones’93 »H/Leland’93 »Related Work –Garbers/Promel/Steger’90, Hagen/Khang’91, Cheng/Wei’91 –Kumar/Karypis’95, etc, etc. l Multigrid Metaphor H/Leland’93 (Chaco) »Popularized by Kumar/Karypis’95 (METIS)

Parallel Computing Sciences Department MOV’01 Multilevel Partitioning l Construct Sequence of Smaller Graphs l Partition Smallest l Project Partition Through Intermediate Levels »Periodically Refine l Why does it work so well? »Refinement on multiple scales (like multigrid) »Key properties preserved on (weighted) coarse graphs –(Weighted) partition sizes –(Weighted) edge cuts »Very fast

Parallel Computing Sciences Department MOV’01 Coarse Problem Construction 1. Find maximal matching 2. Contract matching edges 3. Sum vertex and edge weights Key Properties: Preserves (weighted) partition sizes Preserves (weighted) edge cuts Preserves planarity Related to min-cut algorithm of Karger/Stein’96

Parallel Computing Sciences Department MOV’01 Extension I: Terminal Propagation l Dunlop/Kernighan’85 »Skew partitioning to address constrained vertices l Also useful for parallel computing »Move few vertices when repartitioning »Assign neighboring vertices to near processors »H/Leland/Van Dreissche’96 l Basic idea: »Vertex has gain-like preference to be in particular partition

Parallel Computing Sciences Department MOV’01 Multilevel Terminal Propagation l How to include in multilevel algorithm? l Simple idea: »When vertices get merged, sum preferences »Simple, fast, effective »Coarse problem precisely mimics original

Parallel Computing Sciences Department MOV’01 Extension II: Finding Vertex Separators l Useful for several partitioning applications »E.g. sparse matrix reorderings l One idea: edge separator first, then min cover »Problem: multilevel power on wrong objective l Better to reformulate multilevel method »Find vertex separators directly »H/Rothberg’98

Parallel Computing Sciences Department MOV’01 Multilevel Vertex Separators l Use same coarse constructor »Except edge weights don’t matter l Change local refinement & coarse solve »Can mimic KL/FM l Resulted in improved matrix reordering tool »Techniques now standard

Parallel Computing Sciences Department MOV’01 Extension III: Hypergraph Partitioning Coarse construction »Contract pairs of vertices? »Contract hyperedges? l Traditional refinement methodology l See talk tomorrow by George Karypis

Parallel Computing Sciences Department MOV’01 Envelope Reduction l Reorder rows/columns of symmetric matrix to keep nonzeros near the diagonal

Parallel Computing Sciences Department MOV’01 Graph Formulation l Each row/column is a vertex l Nonzero in (i,j) generates edge e ij l For row i of matrix (vertex i) »Env(i) = max(i-j such that e ij in E) »Envelope =  Env(i) l Find vertex numbering to minimize envelope »NP-Hard

Parallel Computing Sciences Department MOV’01 Status l Highest Quality algorithm is spectral ordering »Sort entries of Fiedler vector (Barnard/Pothen/Simon’95) »Eigenvector calculation is expensive »Fast Sloan (Kumfert/Pothen’97) good compromise l Now multilevel methods are comparable »(Boman/H’96, Hu/Scott’01) l Related ordering problems with VLSI relevance »Optimal Linear Arrangement –Minimize  |i-j| such that e ij in E

Parallel Computing Sciences Department MOV’01 Challenges for Multilevel Envelope Minimization l No precise coarse representation »Can’t express exact objective on coarse problem l No incremental update for envelope metric »I.e. no counterpart of Fiduccia/Mattheyses l Our solution: Use approximate metric »1-sum / minimum linear arrangement »Allows for incremental update –But still not an exact coarse problem »VLSI applications?

Parallel Computing Sciences Department MOV’01 Results l Disappointing for envelope minimization »We never surpassed best competitor –Fast Sloan algorithm »But Hu/Scott’01 succeeded with similar ideas l Better for linear arrangement, but … »Not competitive with Hur/Lillis’99 –Multilevel algorithm with expensive refinement

Parallel Computing Sciences Department MOV’01 Lessons Learned l Good coarse model is the key »Need to encode critical properties of full problem »Progress on coarse instance must help real one »Must allow for efficient refinement methodology »Different objectives require different coarse models l Quality/Runtime tradeoff varies w/ application »Must understand needs of your problem domain »For VLSI, quality is worth waiting for »All aspects of multilevel algorithm are impacted

Parallel Computing Sciences Department MOV’01 Conclusions l Appropriate coarse representation is key »Lots of existing ideas for construction coarse problem –Matching contraction, independent sets, fractional assignment, etc. l Multigrid metaphor provides important insight »We’re not yet fully exploiting multigrid possibilities »Do we have something to offer algebraic multigrid? l Need for CS recognition of multilevel paradigm »Rich, general algorithmic framework, but not in any textbook »Not the same as divide-and-conquer

Parallel Computing Sciences Department MOV’01 Acknowledgements l Shanghua Teng »"Coarsening, Sampling and Smoothing: Elements of the Multilevel Method“ l Rob Leland l Erik Boman l Ed Rothberg l Tammy Kolda l Chuck Alpert l DOE MICS Office