Semi-supervised Affinity Propagation

Slides:



Advertisements
Similar presentations
Clustering. How are we doing on the pass sequence? Pretty good! We can now automatically learn the features needed to track both people But, it sucks.
Advertisements

Perceptron Learning Rule
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Lauritzen-Spiegelhalter Algorithm
Transportation Problem (TP) and Assignment Problem (AP)
1 Lecture 5 Support Vector Machines Large-margin linear classifier Non-separable case The Kernel trick.
Clustering by Passing Messages Between Data Points Brendan J. Frey and Delbert Dueck Science, 2007.
A Probabilistic Framework for Semi-Supervised Clustering
Classification and Decision Boundaries
Non-metric affinity propagation for unsupervised image categorization Delbert Dueck and Brendan J. Frey ICCV 2007.
Branch and Bound Similar to backtracking in generating a search tree and looking for one or more solutions Different in that the “objective” is constrained.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
Active Learning with Support Vector Machines
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. Gerhard Maierbacher Scalable Coding Solutions for Wireless Sensor Networks IT.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
What is Cluster Analysis?
Semi-Supervised Learning D. Zhou, O Bousquet, T. Navin Lan, J. Weston, B. Schokopf J. Weston, B. Schokopf Presents: Tal Babaioff.
Clustering Ram Akella Lecture 6 February 23, & 280I University of California Berkeley Silicon Valley Center/SC.
Clustering Unsupervised learning Generating “classes”
Approximation algorithms for large-scale kernel methods Taher Dameh School of Computing Science Simon Fraser University March 29 th, 2010.
Unsupervised Learning of Categories from Sets of Partially Matching Image Features Kristen Grauman and Trevor Darrel CVPR 2006 Presented By Sovan Biswas.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Radial Basis Function Networks
Fall 2004 TDIDT Learning CS478 - Machine Learning.
SVM by Sequential Minimal Optimization (SMO)
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
Active Learning An example From Xu et al., “Training SpamAssassin with Active Semi- Supervised Learning”
CP Summer School Modelling for Constraint Programming Barbara Smith 2. Implied Constraints, Optimization, Dominance Rules.
Transductive Regression Piloted by Inter-Manifold Relations.
Overview Concept Learning Representation Inductive Learning Hypothesis
Prototype-Driven Learning for Sequence Models Aria Haghighi and Dan Klein University of California Berkeley Slides prepared by Andrew Carlson for the Semi-
1 Network Models Transportation Problem (TP) Distributing any commodity from any group of supply centers, called sources, to any group of receiving.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Clustering by soft-constraint affinity propagation: applications to gene- expression data Michele Leone, Sumedha and Martin Weight Bioinformatics, 2007.
Optimal Reverse Prediction: Linli Xu, Martha White and Dale Schuurmans ICML 2009, Best Overall Paper Honorable Mention A Unified Perspective on Supervised,
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Data Mining: Cluster Analysis This lecture node is modified based on Lecture Notes for Chapter.
Link Analysis Algorithms Page Rank Slides from Stanford CS345, slightly modified.
Machine Learning in Practice Lecture 21 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Data-Driven 3D Voxel Patterns for Object Category Recognition Andrew Sharp.
DATA MINING: CLUSTER ANALYSIS Instructor: Dr. Chun Yu School of Statistics Jiangxi University of Finance and Economics Fall 2015.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
Planning, Acting, and Learning Chapter Contents The Sense/Plan/Act Cycle Approximate Search Learning Heuristic Functions Rewards Instead of Goals.
Clustering Clustering definition: Partition a given set of objects into M groups (clusters) such that the objects of each group are ‘similar’ and ‘different’
Semi-Supervised Clustering
Clustering CSC 600: Data Mining Class 21.
Graphcut Textures:Image and Video Synthesis Using Graph Cuts
Linli Xu Martha White Dale Schuurmans University of Alberta
Chapter - 12 GRAPH MATRICES AND APPLICATIONS.
Constrained Clustering -Semi Supervised Clustering-
Session 7: Face Detection (cont.)
Parallel Programming By J. H. Wang May 2, 2017.
Data Mining K-means Algorithm
cs540 - Fall 2015 (Shavlik©), Lecture 25, Week 14
Object-Graphs for Context-Aware Category Discovery
Overcoming Errors.
Jianping Fan Dept of Computer Science UNC-Charlotte
COSC 4335: Other Classification Techniques
Applied Combinatorics, 4th Ed. Alan Tucker
Support Vector Machines
Gene Tree Estimation Through Affinity Propagation
“Clustering by Passing Messages Between Data Points”
Concave Minimization for Support Vector Machine Classifiers
Compact routing schemes with improved stretch
Label propagation algorithm
Version Space Machine Learning Fall 2018.
Error Correction Coding
Introduction to Machine learning
Presentation transcript:

Semi-supervised Affinity Propagation Inmar Givoni, Brendan Frey, Delbert Dueck PSI group University of Toronto

Affinity Propagation Clustering algorithm that works by finding a set of exemplars (prototypes) in the data and assigning other data points to the exemplars [Frey07] Input: pair-wise similarities (negative squared error), data point preferences (larger = more likely to be an exemplar) Approximate maximization of the sum of similarities to exemplars Mechanism – message passing in a factor graph

Semi-supervised Learning Large amounts of unlabeled training data Some limited amounts of side information Partial labels Equivalence constraints Half moon data

Some Motivating examples

AP with partial labels All points sharing the same label should be in the same cluster. Points with different labels should not be in the same cluster. Imposing constraints Via the similarity matrix Explicit function nodes

Same label constraints Set similarity among all similarly labeled data to be maximal. Propagate to other points (teleportation) Without teleportation, local neighborhoods do not ‘move closer’. e.g. Klein02] S(x1,x2)=0 x1 x2 y2 y1

Different labels Can still do a similar trick and set similarity among all pair-wise differently labeled data to be minimal. But no equivalent notion of anti-teleportation. x1 x2

Adding explicit constraints to account for side-information

Adding explicit constraints to account for side-information

Problems Let’s call all the labeled points portals They induce the ability to teleport… At test time, if we want to determine a label for some new point we need to evaluate its closest exemplar, possibly via all pairs of portals - expensive. Pair-wise not-in-class nodes for each pair of differently labeled points is expensive. Introducing…

Meta-Portals An alternative way of propagating neighborhood information. Meta-portals are ‘dummy’ points, constructed using the similarities of all portals of a certain label. We add N new entries to the similarity matrix, where N is the number of unique labels.

Meta-portals mtp’s can be exemplars. Unlike regular exemplars, mtp’s can be exemplars for other points but choose a different exemplars themselves

These function nodes force the MTP’s to choose other data points as their exemplars. Similarities alone are not enough, since both MTP can choose same exemplars and still have –inf similarities.

Some toy data results

Future work Investigate interplay between modifying similarities and incorporating explicit constraints. Possible tool for user-guided labeling