1 / 27

Enabling Fast Prediction for Ensemble Models on Data Streams

Enabling Fast Prediction for Ensemble Models on Data Streams. Peng Zhang, Jun Li, Peng Wang Byron Gao, Xingquan Zhu, Li Guo Information Security Research Center, Institute of Computing Technology, Chinese Academy of Sciences, China Texas State University-San Marocs, USA

twyla
Download Presentation

Enabling Fast Prediction for Ensemble Models on Data Streams

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Enabling Fast Prediction for Ensemble Models on Data Streams Peng Zhang, Jun Li, Peng Wang Byron Gao, Xingquan Zhu, Li Guo Information Security Research Center, Institute of Computing Technology, Chinese Academy of Sciences, China Texas State University-San Marocs, USA University of Technology, Syndey, Australia

  2. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  3. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  4. Data Stream Classification • Information Security • Applications: Intrusion detection, spam filtering, web page stream monitoring • Example 1. Web page stream monitoring • Two requirements • Classify each incoming web page as accurate as possible. • Classify each incoming web page as fast as possible.

  5. Ensemble Models • Technique • Use a divide-and-conquer manner to build models from continuous data streams • Merits • Scale well, adapt quickly to new concepts, low variance errors, and ease of parallelization. • Family • Ensemble weighted classifiers, incremental classifiers, both classifiers and clusters

  6. Limitation of Ensembles • Low Prediction efficiency. • Prediction usually takes linear time complexity. • Previous works combine a small number of base classifiers in ensemble. • Example 2: Prediction time of ensemble models 830 Processors for 50 classifier ensembles !

  7. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  8. Solutions • Convert ensemble into spatial database • Design a new Ensemble-Tree indexing structure • Prediction becomes search over Ensemble-Tree • Build and update ensemble becomes insertion and deletion. Our solutions

  9. Spatial Indexing for Ensemble • How to convert ensemble into spatial database ? • Example 3: Convert ensemble into spatial database

  10. Spatial Indexing for Ensemble • Why convert ensemble to spatial database ? • Trade space for time. • Example 4. linear vs. sub-linear prediction (b) Spatial database (a) Linear prediction (c) Sub-linear prediction

  11. Ensemble-Tree (E-Tree) Comparisons (classifier_id, weight, pointer) (I, classifier_id, sibling) Parameters M, m • E-Tree is designed for binary classification. • All decision rules are from one class, which is called target class.

  12. Operations on E-Tree • Search • Goal: Predict the label of each incoming stream record x, • Step 1: depth-first traverse to find entries in leaf covering x (i.e., u entries) • Step 2: calculate x’s class label • Example 5. Search • Consider a new record x = (4, 4). • E-Tree achieve 20% improvement in predicting x’s class label.

  13. Operations on E-Tree • Insertion • Goal: Integrates new classifiers into ensemble • Step 1: Insert the classifier into the table structure • Step 2: For each rule R in the classifier • search the leaf node that contains R • If the leaf node is full, split the node; otherwise, add in directly. • Repeat until all nodes meet the [m, M] constraint. • Example 6. Insertion • We insert the five classifiers one by one. Suppose parameters M=3, and m=1.

  14. Operations on E-Tree • Deletion • Goal: Discards outdated classifiers from ensemble • Step 1: Delete an outdated classifier from the table. • Step 2: Delete all entries associated with the classifier • Step 3: After deletion, if a node has less than m entries, then delete the leaf node, and re-insert into the tree. • Step 4: Repeat until all nodes meet the [m, M] constraint.

  15. Ensemble Learning with E-Trees • The architecture • online prediction: search operation • Sideway training and updating: Insertion and deletion operations

  16. Theoretical Analysis of E-Trees • For each incoming record x, the ideal situation is that all target decision rules that cover x are located in one single leaf. (1) What's the probability of x's target decision rules being located in one single leaf node? (2) In the ideal case, what’s the worst prediction time? • The worst case is that the target decision rules are uniformly distributed across m leaf nodes. (3) In the worst case, what’s the worst prediction time?

  17. Theoretical Analysis of E-Trees Sublinear time

  18. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  19. Experiments • Data sets • Synthetic data • Real world data • Comparison methods • Global E-Tree, Local E-Tree • Global-Ensemble, Local-Ensemble Decision space 1 0 1

  20. Experiments • Prediction time

  21. Experiments • Memory cost

  22. Experiments • Four methods comparisons • LE-tree performs better than L-Ensemble with respect to prediction time. • GE-tree performs better than G-Ensemble with respect to prediction time. • LE-tree achieves the same prediction accuracy as L-Ensemble. • GE-tree achieves the same prediction accuracy as G-Ensemble. E-Trees can reduce Ensemble model’s prediction time, meanwhile, E-Tree can achieve the same prediction accuracy as ensemble.

  23. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  24. Related Work • Stream classification • Incremental learning, ensemble learning • Stream indexing • Boolean expression indexing in Publish/subscribe systems • Spatial indexing • R-Tree, R*-Tree, R+-Tree, X-Tree, … • Ensemble Pruning • Concept drifting

  25. Outline • Motivation • Solutions • Experiments • Related work • Conclusion

  26. Conclusion • Contributions • Identify and address the prediction efficiency problem for ensemble models on data streams • Convert ensemble model to spatial database, and propose a new type of spatial indexing Ensemble-Tree to achieve sub-linear prediction time. • Future work • For time-critical data mining tasks. • Index SVM and other classification models.

  27. Source codehttp://streammining.org Thank you ! Questions?

More Related