Post Silicon Test Optimization Ron Zeira 13.7.11.

Slides:



Advertisements
Similar presentations
Clustering k-mean clustering Genome 559: Introduction to Statistical and Computational Genomics Elhanan Borenstein.
Advertisements

K-means Clustering Given a data point v and a set of points X,
Clustering.
Albert Gatt Corpora and Statistical Methods Lecture 13.
Data Mining Techniques: Clustering
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Introduction to Bioinformatics
Statistics for Marketing & Consumer Research Copyright © Mario Mazzocchi 1 Cluster Analysis (from Chapter 12)
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
Analyzing System Logs: A New View of What's Important Sivan Sabato Elad Yom-Tov Aviad Tsherniak Saharon Rosset IBM Research SysML07 (Second Workshop on.
Introduction to Bioinformatics Algorithms Clustering.
Cluster Analysis (1).
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Introduction to Bioinformatics - Tutorial no. 12
What is Cluster Analysis?
K-means Clustering. What is clustering? Why would we want to cluster? How would you determine clusters? How can you do this efficiently?
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
Clustering. What is clustering? Grouping similar objects together and keeping dissimilar objects apart. In Information Retrieval, the cluster hypothesis.
Introduction to Bioinformatics Algorithms Clustering and Microarray Analysis.
Clustering Unsupervised learning Generating “classes”
Evaluating Performance for Data Mining Techniques
Gene expression & Clustering (Chapter 10)
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
A Clustering Algorithm based on Graph Connectivity Balakrishna Thiagarajan Computer Science and Engineering State University of New York at Buffalo.
Clustering Algorithms k-means Hierarchic Agglomerative Clustering (HAC) …. BIRCH Association Rule Hypergraph Partitioning (ARHP) Categorical clustering.
1 Gene Ontology Javier Cabrera. 2 Outline Goal: How to identify biological processes or biochemical pathways that are changed by treatment.Goal: How to.
CLUSTERING. Overview Definition of Clustering Existing clustering methods Clustering examples.
Christoph F. Eick Questions and Topics Review Dec. 6, Compare AGNES /Hierarchical clustering with K-means; what are the main differences? 2 Compute.
Microarray Data Analysis (Lecture for CS498-CXZ Algorithms in Bioinformatics) Oct 13, 2005 ChengXiang Zhai Department of Computer Science University of.
CS654: Digital Image Analysis
Quantitative analysis of 2D gels Generalities. Applications Mutant / wild type Physiological conditions Tissue specific expression Disease / normal state.
Clustering.
By Timofey Shulepov Clustering Algorithms. Clustering - main features  Clustering – a data mining technique  Def.: Classification of objects into sets.
K-Means Algorithm Each cluster is represented by the mean value of the objects in the cluster Input: set of objects (n), no of clusters (k) Output:
Selecting Diverse Sets of Compounds C371 Fall 2004.
Clustering Clustering is a technique for finding similarity groups in data, called clusters. I.e., it groups data instances that are similar to (near)
Gene expression & Clustering. Determining gene function Sequence comparison tells us if a gene is similar to another gene, e.g., in a new species –Dynamic.
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Machine Learning Queens College Lecture 7: Clustering.
Definition Finding groups of objects such that the objects in a group will be similar (or related) to one another and different from (or unrelated to)
Clustering Patrice Koehl Department of Biological Sciences National University of Singapore
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Network Partition –Finding modules of the network. Graph Clustering –Partition graphs according to the connectivity. –Nodes within a cluster is highly.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Clustering Approaches Ka-Lok Ng Department of Bioinformatics Asia University.
CURE: An Efficient Clustering Algorithm for Large Databases Authors: Sudipto Guha, Rajeev Rastogi, Kyuseok Shim Presentation by: Vuk Malbasa For CIS664.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
Unsupervised Classification
Clustering [Idea only, Chapter 10.1, 10.2, 10.4].
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Data Mining and Text Mining. The Standard Data Mining process.
CLUSTER ANALYSIS. Cluster Analysis  Cluster analysis is a major technique for classifying a ‘mountain’ of information into manageable meaningful piles.
Unsupervised Learning: Clustering
Unsupervised Learning: Clustering
Clustering Patrice Koehl Department of Biological Sciences
Machine Learning Clustering: K-means Supervised Learning
Data Mining K-means Algorithm
Dr. Unnikrishnan P.C. Professor, EEE
Information Organization: Clustering
KMeans Clustering on Hadoop Fall 2013 Elke A. Rundensteiner
Data Mining 資料探勘 分群分析 (Cluster Analysis) Min-Yuh Day 戴敏育
Clustering.
Text Categorization Berlin Chen 2003 Reference:
Clustering Techniques
Data Mining CSCI 307, Spring 2019 Lecture 24
Clustering.
Presentation transcript:

Post Silicon Test Optimization Ron Zeira

Background Post-Si validation is the validation of the real chip on the board. It takes many resources, both machine and human. Therefore it is important to keep less tests in the suite, but these tests must be efficient.

DB Events Tests s1 s2 s3 s4 s5 s6 Test 1 s1 s2 s3 s4 s5 s6 Test 2 Test 3 s1 s2

DB Take the maximum hit over seeds. Filter results below threshold. Results in a 722X108 testXevent matrix 11.8% full. A test has 3 sets with the system elements, configurations and modifications it ran with.

Event Covering techniques Single event covering: ◦ Set cover. ◦ Dominating set. Event pairs covering: ◦ Pair set cover. ◦ Pair dominating set. Undominated tests.

Test X Event matrix Tests Events

Test clustering The goal is to find groups of similar tests. ◦ First attempts with expander. ◦ Similarity measure. ◦ Binary K-Means. ◦ Other binary methods clustering.

Clustering with expander Tests Events

Hit count drawbacks Pearson correlation\Euclidian distance consider sparse vectors similar. Hit counts are deceiving. Normaliztion.

Binary similarity measure Consider tests as binary vectors or sets. Hamming distance – doesn’t differ between 0-0 and 1-1 Jaccard coefficient: ◦ Pro - Prefers 1’s over 0’s. ◦ Con – usually underestimates similarity.

Binary similarity measure Geometric mean/dot product/cosine/Ochiai: Arithmetic mean: Geometric mean ≤ arithmetic mean

Binary similarity measure ArithmeticGeometricJaccard ααα/(2-α) Undervalued overlap |v1|=|v2|=k |v1 ∩ v2| = αk (1+ α)/2 Sqrt(α)α Undervalued part v1 ⊆ v2 |v1 ∩ v2| =|v1|= α|v2| Similarities: Jaccard ≤(?) Geometric mean ≤ arithmetic mean

Test Clustering (Jaccard similarity) Hierarchical cluster divided to 8 clusters. Done with R

Binary K-means Choose initial solution. While not converge ◦ Move a test to the cluster it is most similar to How to calculate the dissimilarity between a cluster to a single set using the binary dissimilarity measures?

Binary K-means Test 2 cluster similarity: 1.Calculate binary centroid. Then check similarity. 2.Use the average similarity to the cluster. 3.Use the minimum/maximum similarity to the cluster.

Binary K-means Choose initial k representatives: ◦ Choose disjoint tests as representatives. ◦ Choose tests with some overlaps.

Evaluating clustering High homogeneity and low separation (function of the similarity). Average Silhouette: how similar each test to its own cluster than to the “closest” cluster. Cluster distribution.

Click CLICK is a graph-based algorithm for clustering. Gets a homogeneity threshold. Run click with dot product similarity. Allows outliers – reflect a unique behavior.

Cluster common features Similar tests (outputs) should have similar configurations (inputs)? Find dominant configs/elements in each cluster using hyper-geometric p-val. Look on configuration “homogeneity”.

Cluster common features Random partition hom Random partition # Config homogeneity P-val & >50%#Configs p- val > 1e Cluster1 (141) Cluster 2 (136) Cluster 3 (54) Cluster 4 (51) Cluster 5 (38) Cluster 6 (33) Singletons (269)

Common features – open issues Compare similarities matrixes according to event and features. Compare cluster solutions according to the features. Given a clustering solution analyze the feature’s role.

Choose the “best” tests to run Do until size is met: ◦ Select the “best” test to add or the “worst” test to remove. Good/Bad test? ◦ Similarity. ◦ Coverage. ◦ Cluster.

Evaluate a test subset Number of event multi-covered. Minimal event multi-covered. Minimal homogeneity. Feature based.

“Farthest” first generalization Start with arbitrary or known subset (cover). At each iteration add the most dissimilar (remove most similar) test to the current selection. Dis/Similar? ◦ Average test to subset dissimilarity. ◦ Minimal test to subset dissimilarity.

Coverage based Start with arbitrary or known subset (cover). At each iteration find the event least covered and add a test that cover it. Similar to set multi-cover.

Cluster based Add singletons to cover. Choose arbitrary cluster according to size, then choose a test from it. Choose the cluster according to the centroid.

Min event cover Average event cover Homogen eity Undomin ated left Event pair cover percentag e Event cover percentag e Add farthest first Remove closest first Add min event Cluster based random Choose 400 tests with no initial cover

Config homoge neity Config in use SE homoge neity SE in use Min event cover Average event cover Homoge neity % % Add farthest first % % % % Remove closest first % % % % Add min event % % random Choose 400 tests with initial 291 un-dominated