1 / 50

Dynamic Integration of Virtual Predictors

Dynamic Integration of Virtual Predictors. Vagan Terziyan University of Jyvaskyla, Finland e-mail: vagan@it.jyu.fi http://www.cs.jyu.fi/ai/vagan/index.html. Discovering Knowledge from Data - one of the basic abilities of an intelligent agent. Data. Knowledge. Basic Reference.

juanjoyce
Download Presentation

Dynamic Integration of Virtual Predictors

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Dynamic Integration of Virtual Predictors Vagan Terziyan University of Jyvaskyla, Finland e-mail: vagan@it.jyu.fi http://www.cs.jyu.fi/ai/vagan/index.html

  2. Discovering Knowledge from Data - one of the basic abilities of an intelligent agent Data Knowledge

  3. Basic Reference Terziyan V., Dynamic Integration of Virtual Predictors, In: L.I. Kuncheva, F. Steimann, C. Haefke, M. Aladjem, V. Novak (Eds), Proceedings of the International ICSC Congress on Computational Intelligence: Methods and Applications- CIMA'2001, Bangor, Wales, UK, June 19 - 22, 2001, ICSC Academic Press, Canada/The Netherlands, pp. 463-469.

  4. Acknowledgements Academy of Finland Project (1999): Dynamic Integration of Classification Algorithms Information Technology Research Institute (University of Jyvaskyla): Customer-oriented research and development in Information Technology http://www.titu.jyu.fi/eindex.html Multimeetmobile (MMM) Project (2000-2001): Location-Based Service System and Transaction Management in Mobile Electronic Commerce http://www.cs.jyu.fi/~mmm

  5. Contents • The problem • Virtual Predictor • Classification Team • Team Direction • Dynamic Selection of Classification Team • Implementation for Mobile e-Commerce • Conclusion

  6. The Problem:Knowledge Discovery • Knowledge discovery in databases (KDD) is a combination of data warehousing, decision support, and data mining and it is an innovative new approach to information management. • KDD is an emerging area that considers the process of finding previously unknown and potentially interesting patterns and relations in large databases*. • __________________________________________________________________________________________________________________________________* Fayyad, U., Piatetsky-Shapiro, G., Smyth, P., Uthurusamy, R., Advances in Knowledge Discovery and Data Mining, AAAI/MIT Press, 1996.

  7. Classification Problem m classes, n training observations, p object features Given: n training pairs (xi, yi) with xiÎRp and yiÎ{1,…,m} denoting class membership Goal: given: new x0 select classifier for x0 predict class y0

  8. The Research Problem During the past several years, in a variety of application domains, researchers in machine learning, computational learning theory, pattern recognition and statistics have tried to combine efforts to learn how to create and combine an ensemble of classifiers. The primary goal of combining several classifiers is to obtain a more accurate prediction than can be obtained from any single classifier alone.

  9. Approaches to Integrate Multiple Classifiers Integrating Multiple Classifiers Combination Selection Decontextualization Local (“Virtual” Classifier) Local(Dynamic) Global (Static) Global (Voting-Type)

  10. Inductive learning with integration of predictors Learning Environment Sample Instances Predictors/Classifiers P1 P2 ... Pn yt

  11. Virtual Classifier Virtual Classifier is a group of seven cooperative agents: TC - Team Collector TM - Training Manager TP - Team Predictor TI - Team Integrator FS - Feature Selector DE - Distance Evaluator CL - Classification Processor

  12. Classification Team: Feature Selector FS - Feature Selector

  13. Feature Selection • Feature selection methods try to pick a subset of features that are relevant to the target concept; • Each of these methods has its strengths and weaknesses based on data types and domain characteristics; • The choice of a feature selection method depends on various data set characteristics: (i) data types, (ii) data size, and (iii) noise.

  14. Classification of feature selection methods [Dash and Liu, 1997]

  15. Feature Selector:to find the minimally sized feature subset that is sufficient for correct classification of the instance Sample Instances Sample Instances Feature Selector

  16. Classification Team: Distance Evaluator DE - Distance Evaluator

  17. Use of distance evaluation • Distance between instances is useful to recognize nearest neighborhood of any classified instance; • Distance between classes is useful to define the misclassification error; • Distance between classifiers is useful to evaluate weights of every classifier for their further integration.

  18. Well known distance functions [ Wilson & Martinez 1997]

  19. Distance between Two Instances with Heterogeneous Attributes (e.g. Profiles) where: d (“red”, “yellow”) = 1 d (15°, 25°) = 10°/((+50°)-(-50°)) = 0.1

  20. Distance Evaluator:to measure distance between instances based on their numerical or nominal attribute values Distance Evaluator

  21. Classification Team: Classification Processor CL - Classification Processor

  22. Classification Processor: to predict class for a new instance based on its selected features and its location relatively to sample instances Feature Selector Sample Instances Classification Processor Distance Evaluator

  23. Team Instructors:Team Collector TC - Team Collector completes Classification Teams for training

  24. Team Collector - completes classification teams for future training Distance Evaluation functions Classification rules Feature Selection methods Team Collector FSi DEj CLk

  25. Team Instructors:Training Manager TM - Training Manager trains all completed teams on sample instances

  26. Training Manager - trains all completed teams on sample instances Training Manager CLk1 FSi1 DEj1 Sample Instances Sample Metadata CLk2 FSi2 DEj2 CLkn FSin DEjn Classification Teams

  27. Team Instructors:Team Predictor TP - Team Predictor predicts weights for every classification team in certain location

  28. Team Predictor - predicts weights for every classification team in certain location Predicted weights of classification teams Location Team Predictor: e.g. WNN algorithm Sample Metadata

  29. Team Predictor - predicts weights for every classification team in certain location Sample metadata NN2 <w21, w22,…, w2n> <w31, w32,…, w3n> NN3 d2 NN1 <w11, w12,…, w1n> d3 d1 Pi dmax <wi1, wi2,…, win> NN4 <w41, w42,…, w4n> wij= F(w1j, d1, w2j, d2, w3j, d3, dmax)

  30. Weighting Neighbors Example

  31. Team Prediction: Locality assumption Each team has certain subdomains in the space of instance attributes, where it is more reliable than the others; This assumption is supported by the experiences, that classifiers usually work well not only in certain points of the domain space, but in certain subareas of the domain space [Quinlan, 1993]; If a team does not work well with the instances near a new instance, then it is quite probable that it will not work well with this new instance also.

  32. Team Instructors:Team Integrator TI - Team Integrator produces classification result for a new instance by integrating appropriate outcomes of learned teams

  33. Team integrator - produces classification result for a new instance by integrating appropriate outcomes of learned teams Weights of classification teams in the location of a new instance New instance CLk1 yt1 FSi1 DEj1 CLk2 yt2 FSi2 DEj2 yt Team Integrator CLkn yt1 FSin DEjn Classification teams

  34. Dynamic Selection of the Team: Penalty Kick Example x4 team selected for the kick x2 x1 x3

  35. Simple case:static or dynamic selection of a classification team from two Assume that we have two different classification teams and they have been learned on a same sample set with n instances. Let the first team classifies correctly m1, and the second one m2sample instances respectively. We consider two possible cases to select the best team for further classification: a static selection case and a dynamic selection case.

  36. Static Selection • Static selection means that we try all teams on a sample set and for further classification select one, which achieved the best classification accuracy among others for the whole sample set. Thus we select a team only once and then use it to classify all new domain instances.

  37. Dynamic Selection • Dynamic selection means that the team is being selected for every new instance separately depending on where this instance is located. If it has been predicted that certain team can better classify this new instance than other teams, then this team is used to classify this new instance. In such case we say that the new instance belongs to the “competence area” of that classification team.

  38. Theorem • The average classification accuracy in the case of (dynamic) selection of a classification team for every instance is expected to be not worse than the one in the case of (static) selection for the whole domain. • The accuracy of these two cases can be equal if and only if : where k is amount of same instances correctly classified by both teams

  39. “Competence areas ” of classification teams in dynamic selection n instances m2 instances m1 instances k instances

  40. M-Commerce LBS systemhttp://www.cs.jyu.fi/~mmm In the framework of the Multi Meet Mobile (MMM) project at the University of Jyväskylä, a LBS pilot system, MMM Location-based Service system (MLS), has been developed. MLS is a general LBS system for mobile users, offering map and navigation across multiple geographically distributed services accompanied with access to location-based information through the map on terminal’s screen. MLS is based on Java, XML and uses dynamic selection of services for customers based on their profile and location.

  41. Positioning Service Geographical, spatial data Mobile network Location-Based Service Personal Trusted Device Location-based data: (1) services database (access history); (2) customers database (profiles) Architecture of LBS system

  42. Positioning systems Cellular network based positioning Satellite-based positioning

  43. Opening a Connection to Location Service

  44. Request and Receive Map from the Location Service

  45. Selecting Point of Intereston the Map

  46. Receiving Information Content Related to Point of Interest

  47. Sample from the location-based services’ access history Mobile customer description Ordered service

  48. Adaptive interface for MLS client Only predicted services, for the customer with known profile and location, will be delivered from MLS and displayed at the mobile terminal screen as clickable “points of interest”

  49. Route-based personalization [Katasonov, 2001] Dynamic Perspective Static Perspective

  50. Conclusion • Knowledge discovery with an ensemble of classifiers is known to be more accurate than with any classifier alone [e.g. Dietterich, 1997]. • If a classifier somehow consists of certain feature selection algorithm, distance evaluation function and classification rule, then why not to consider these parts also as ensembles making a classifier itself more flexible? • We expect that classification teams completed from different feature selection, distance evaluation, and classification methods will be more accurate than any ensemble of known classifiers alone, and we focus our research and implementation on this assumption.

More Related