1 / 29

Approximating Sensor Network Queries Using In-Network Summaries

Approximating Sensor Network Queries Using In-Network Summaries. Alexandra Meliou Carlos Guestrin Joseph Hellerstein. Approximate Answer Queries. Approximate representation of the world: Discrete locations Lossy communication Noisy measurements

berg
Download Presentation

Approximating Sensor Network Queries Using In-Network Summaries

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Approximating Sensor Network Queries Using In-Network Summaries Alexandra Meliou Carlos Guestrin Joseph Hellerstein

  2. Approximate Answer Queries • Approximate representation of the world: • Discrete locations • Lossy communication • Noisy measurements • Applications do not expect accurate values (tolerance to noise) • Example: • Return the temperature at all locations ±1C, with 95% confidence • Query Satisfaction: • On expectation the requested portion of sensor values lies within the error range

  3. In-network Decisions Query Use in-network models to make routing decisions No centralized planning

  4. In-network Summaries Spanning tree T(V,E’) + Models Mv for all nodes v Mv represents the whole subtree rooted at v.

  5. Model Complexity • Gaussian distributions at the leaves: • good for modeling individual node measurements Need for compression

  6. Talk “outline” Compression In-network summaries Construction Traversal

  7. Collapsing Gaussian Mixtures • Compress an m-size mixture to a k-size mixture. • Look at simple case (k=1) • Minimize KL-divergence? “Fake” mass

  8. Quality of Compression Depends on query workload Query with acceptable error window W’<W Query with acceptable error window W

  9. Compression Accurate mass inside interval No guarantee on the tails

  10. Talk “outline” Compression In-network summaries Construction Traversal

  11. Query Satisfaction • A response R={r1…rn} satisfies query Q(w,δ) if: • In expectation the values of at least δn nodes lie within [ri-w,ri+w] Q In-network summary Within error bounds [r1, r2, r3, r4, r5, r6, r7, r8, r9, r10] R

  12. Optimal Traversal • Given: tree and models • Find: subtree such that [μleaves] response Can be computed with Dynamic Programming

  13. Greedy Traversal • If local model satisfies • Return μ • Else descend to child node More conservative solution: enforces query satisfiability on every subtree instead of the whole tree

  14. Traversal Evaluation

  15. Talk “outline” Compression In-network summaries Construction Traversal

  16. Optimal Tree Construction • Given a structure, we know how to build the models • But how do we pick the structure?

  17. Traversal = cut Theorem: In a fixed fanout tree, the cost of the traversal is where |C| is the size of the cut, and F the fanout Intuition: minimize cut size Group nodes into a minimum number of groups which satisfy the query constraints Clustering problem

  18. Optimal Clustering • Given a query Q(w,δ), optimal clustering is NP-hard • Related to the Group Steiner Tree Problem • Greedy algorithm with factor log(n) approximation • Greedily pick max size cluster • Issue: does not enforce connectivity of clusters

  19. Greedy Clustering • Include extra nodes to enforce connectivity • Augment clusters only with accessible nodes (losing the logn guarantee)

  20. Clustering comparison • 2 distributed clustering algorithms are compared to the centralized greedy clustering

  21. Talk “outline” Compression Enriched models In-network summaries Construction Traversal

  22. Enriched models • Support more complex models • k-mixtures • Compress to a k-size mixture instead of a SGM • Virtual nodes • Every component of the k-size mixture is stored as a separate “virtual node” • SGMs on multiple windows • Maintain additional SGMs for different window sizes • More space, more expensive model updates (SGM = Single Gaussian Model)

  23. Evaluation of enriched models SGM surprisingly effective in representing the underlying data

  24. Talk “outline” Sensitivity analysis Compression In-network summaries Construction Traversal

  25. Tree Construction Parameters and Effect on Performance • Confidence • Performance for workloads of different confidence than the hierarchy design • Error window • Broader vs narrower ranges of window sizes • Assignment of windows across tree levels • Temporal changes • How often should the models be updated

  26. Confidence Workload of 0.95 confidence Design confidence does not have a big impact on performance

  27. Error windows A wide range is not always better, because it forces the traversal of more levels

  28. Model Updates

  29. Compression Conclusions Enriched models In-network summaries Traversal Construction • Analyzed compression schemes for in-network summaries • Evaluated summary traversal • Studied optimal hierarchy construction • Studied increased complexity models • Showed that simple SGM are sufficient • Analyzed the effect on efficiency of various parameters Sensitivity analysis

More Related