Download presentation
Presentation is loading. Please wait.
1
Approximating Sensor Network Queries Using In-Network Summaries Alexandra Meliou Carlos Guestrin Joseph Hellerstein
2
Approximate Answer Queries Approximate representation of the world: Discrete locations Lossy communication Noisy measurements Applications do not expect accurate values (tolerance to noise) Example: Return the temperature at all locations ±1C, with 95% confidence Query Satisfaction: On expectation the requested portion of sensor values lies within the error range
3
In-network Decisions Query Use in-network models to make routing decisions No centralized planning
4
In-network Summaries Spanning tree T(V,E’) + Models M v for all nodes v M v represents the whole subtree rooted at v.
5
Model Complexity Need for compression Gaussian distributions at the leaves: good for modeling individual node measurements Gaussian distributions at the leaves: good for modeling individual node measurements
6
Talk “outline” Compression Traversal Construction In-network summaries
7
Collapsing Gaussian Mixtures Compress an m-size mixture to a k-size mixture. Look at simple case (k=1) Minimize KL- divergence? “Fake” mass
8
Quality of Compression Depends on query workload Query with acceptable error window W Query with acceptable error window W’<W
9
Compression Accurate mass inside interval No guarantee on the tails
10
Talk “outline” Compression Traversal Construction In-network summaries
11
Query Satisfaction A response R={r 1 …r n } satisfies query Q(w,δ) if: In expectation the values of at least δn nodes lie within [r i -w,r i +w] In-network summary Q R [r 1, r 2, r 3, r 4, r 5, r 6, r 7, r 8, r 9, r 10 ] Within error bounds
12
Optimal Traversal Given: tree and models Find: subtree such that Can be computed with Dynamic Programming response [μ leaves ]
13
Greedy Traversal If local model satisfies Return μ Else descend to child node More conservative solution: enforces query satisfiability on every subtree instead of the whole tree More conservative solution: enforces query satisfiability on every subtree instead of the whole tree
14
Traversal Evaluation
15
Talk “outline” Compression Traversal Construction In-network summaries
16
Optimal Tree Construction Given a structure, we know how to build the models But how do we pick the structure?
17
Traversal = cut Theorem: In a fixed fanout tree, the cost of the traversal is where |C| is the size of the cut, and F the fanout Theorem: In a fixed fanout tree, the cost of the traversal is where |C| is the size of the cut, and F the fanout Intuition: minimize cut size Group nodes into a minimum number of groups which satisfy the query constraints Clustering problem
18
Optimal Clustering Given a query Q(w,δ), optimal clustering is NP-hard Related to the Group Steiner Tree Problem Greedy algorithm with factor log(n) approximation Greedily pick max size cluster Issue: does not enforce connectivity of clusters
19
Greedy Clustering Include extra nodes to enforce connectivity Augment clusters only with accessible nodes (losing the logn guarantee)
20
Clustering comparison 2 distributed clustering algorithms are compared to the centralized greedy clustering
21
Talk “outline” Compression Traversal Construction In-network summaries Enriched models
22
Support more complex models k-mixtures Compress to a k-size mixture instead of a SGM Virtual nodes Every component of the k-size mixture is stored as a separate “virtual node” SGMs on multiple windows Maintain additional SGMs for different window sizes More space, more expensive model updates (SGM = Single Gaussian Model)
23
Evaluation of enriched models SGM surprisingly effective in representing the underlying data
24
Sensitivity analysis Talk “outline” Compression Traversal Construction In-network summaries
25
Tree Construction Parameters and Effect on Performance Confidence Performance for workloads of different confidence than the hierarchy design Error window Broader vs narrower ranges of window sizes Assignment of windows across tree levels Temporal changes How often should the models be updated
26
Confidence Workload of 0.95 confidence Design confidence does not have a big impact on performance
27
Error windows A wide range is not always better, because it forces the traversal of more levels
28
Model Updates
29
Sensitivity analysis Conclusions Analyzed compression schemes for in-network summaries Evaluated summary traversal Studied optimal hierarchy construction Studied increased complexity models Showed that simple SGM are sufficient Analyzed the effect on efficiency of various parameters Compression Traversal Construction In-network summaries Enriched models
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.