Download presentation
Presentation is loading. Please wait.
1
Solving the Multiple-Instance Problem with Axis-Parallel Rectangles By Thomas G. Dietterich, Richard H. Lathrop, and Tomás Lozano-Pérez Appeared in Artificial Intelligence, Volume 89, Issue1-2, (January 1997), Pages: 31 – 71. Presented by Shuiwang Ji Part of this slides are from the original paper and from Dr. Yixin Chen.
2
The key-lock problem Problem statements: 1.There is a keyed lock on the door to the supply room; 2.Each staff member has a key chain containing several keys; 3.One key on each key chain can open the supply room door; 4.For some staff members, their supply key may open one or more other doors; Positive bag Negative bag
3
The key-lock problem Question: You are a lock smith and you are attempting to infer the most general required shape that a key must have in order to open the supply room door.
4
Drug activity prediction Problem statements: A good drug molecule will bind very tightly to the desired binding site; The input object is a molecule, and the observed result is “bind” or “not bind”; Conformation determines bind or not; Each drug molecule has many conformations;
5
Drug activity prediction Problem statements: A good drug molecule will bind very tightly to the desired binding site; The input object is a molecule, and the observed result is “bind” or “not bind”; Conformation determines bind or not; Each drug molecule has many conformations; Positive bag Negative bag
6
Drug activity prediction Question: The goal is to infer the most general shape required for binding to the binding site.
7
Supervised learning
8
Multiple-instance learning Key chain Molecule Key Conformations Open? Bind? bag
9
Formal problem definition Supervised learning: Multiple-instance learning:
10
Ray-based representation A set of 162 rays emanating from the origin; Each feature value is the distance from the origin to the molecular surface.
11
Three classes of algorithms
12
Standard APR algorithm 1 1 1 1 1 1 5 4 2 2 2 2 2 2 3 3 3 3 3 4 4 4 4 5 5 5 5 x1x1 x2x2 2 For each negative instance, count the number of instances that must be excluded from the APR in order to exclude the negative instance. Greedy algorithm: Eliminate the negative instance that requires eliminating the fewest positive instances
13
Standard APR algorithm Only the NUMBER of positive instances is considered when eliminating negative instances; The resulting APR may not cover at least one instance from all positive bags; The cost of eliminating each positive instance should be different.
14
Outside-in APR algorithm 1 1 1 1 1 1 5 4 2 2 2 2 2 2 3 3 3 3 3 4 4 4 4 5 5 5 5 x1x1 x2x2 2 Consider excluding positive instances that are expendable in the sense that every positive molecule still has at least one positive instance Greedy algorithm: define a cost function for the elimination 4
15
Outside-in APR algorithm 1 1 1 1 1 1 5 4 2 2 2 2 2 2 3 3 3 3 3 4 4 4 4 5 5 5 5 x1x1 x2x2 2 Consider excluding positive instances that are expendable in the sense that every positive molecule still has at least one positive instance Greedy algorithm: define a cost function for the elimination 4
16
Outside-in APR algorithm The cost of excluding a positive instance of molecule depends on the other not yet excluded positive instances of the same molecule; A cost function based on the density estimation of the positive instances was proposed; The employed density estimation is expensive to compute.
17
Inside-out APR algorithm 1 1 1 1 1 1 5 4 2 2 2 2 2 2 3 3 3 3 3 4 4 4 4 5 5 5 5 x1x1 x2x2 2 Choose an initial seed positive instance Grow the APR by identifying the positive instance that when added to the APR would least increase its size.
18
Grow: An algorithm for growing an APR with tight bounds along a specified set of features Inside-out APR algorithm Discrimination: An algorithm for choosing a set of discriminating features Expand: An algorithm for expanding the bounds of an APR to improve its generalization ability
19
Inside-out APR algorithm (grow) The size of an APR is the sum of the widths of all of its bounds. Greedy: Identify the positive instance of a not yet covered positive molecule that would least increase its size; 1,2,…,d-1,d, d+1 Backfitting: After making the d-th decision, all previous d- 1 decisions are revisited until no changes. 1,2,…,d-1, d, d+1
20
Inside-out APR algorithm (discrimination) A feature strongly discriminates against a negative instance if that instance is far outside of the bound of this feature; In each step, choose a feature that strongly discriminates against the largest number of negative instances; Repeat until all negative instances are excluded.
21
Inside-out APR algorithm (expand) APR resulting from the first two steps is too tight since it is designed to cover at least one positive instance of each positive bag; Apply kernel density estimation to estimate the probability that a positive instance will satisfy the bounds on that feature; Expand the bounds so that with high probability, new positive instance will fall inside the APR.
22
Performance of iterated discrimination on artificial data set
23
Performance of iterated discrimination on musk data set 1
24
Performance of iterated discrimination on musk data set 2
25
Thank you!
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.