1 / 24

Data Mining

Data Mining. CMPT 455/826 - Week 10, Day 2. A Methodology for Evaluating and Selecting Data Mining Software. (based on Collier). Evaluation Categories. Performance is the ability to handle a variety of data sources in an efficient manner Functionality

Download Presentation

Data Mining

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Mining CMPT 455/826 - Week 10, Day 2 Jan-Apr 2009 – w10d2

  2. A Methodology for Evaluating and Selecting Data Mining Software (based on Collier) Jan-Apr 2009 – w10d2

  3. Evaluation Categories • Performance • is the ability to handle a variety of data sources in an efficient manner • Functionality • is the inclusion of a variety of capabilities, techniques, and methodologies for data mining • Usability • is accommodation of different levels and types of users without loss of functionality or usefulness • Ancillary Task Support • allows the user to perform the variety of data cleansing, manipulation, transformation, visualization and other tasks that support data mining Jan-Apr 2009 – w10d2

  4. Methodology • Tool pre-screening • Identify Additional Selection Criteria • Weight Selection Criteria • Tool Scoring • Score Evaluation • Tool Selection Jan-Apr 2009 – w10d2

  5. Introduction to Data Mining from Discovering Knowledge in Data: An Introduction to Data Mining by Daniel T. Larose Jan-Apr 2009 – w10d2

  6. Need For Human Direction Of Data Mining • Many software vendors market their analytical software as being plug-and-play out of-the-box applications that will provide solutions to otherwise intractable problems without the need for human supervision or interaction. Some early definitions of data mining followed this focus on automation. • Humans need to be actively involved at every phase of the data mining process. • Rather than asking where humans fit into data mining, we should instead inquire about how we may design data mining into the very human process of problem solving. Jan-Apr 2009 – w10d2

  7. CRISP–DM: The Six Phases • Business understanding phase • Data understanding phase • Data preparation phase • Modeling phase • Evaluation phase • Deployment phase Jan-Apr 2009 – w10d2

  8. CRISP–DM: Modeling phase • Select and apply appropriate modeling techniques • Calibrate model settings to optimize results • Remember that often, several different techniques may be used for the same data mining problem • If necessary, loop back to the data preparation phase to bring the form of the data into line with the specific requirements of a particular data mining technique Jan-Apr 2009 – w10d2

  9. CRISP–DM: Evaluation phase • Evaluate the one or more models delivered in the modeling phase • for quality and effectiveness before deploying them for use in the field. • Determine whether the model in fact achieves the objectives • set for it in the first phase. • Establish if something has not been accounted for sufficiently • some important facet of the business or research problem • Come to a decision • regarding use of the data mining results. Jan-Apr 2009 – w10d2

  10. Fallacies Of Data Mining • Fallacy 1. There are data mining tools that we can turn loose on our data repositories and use to find answers to our problems. • Reality. There are no automatic data mining tools that will solve your problems mechanically "while you wait." Rather, data mining is a process, as we have seen above. CRISP–DM is one method for fitting the data mining process into the overall business or research plan of action. Jan-Apr 2009 – w10d2

  11. Fallacies Of Data Mining • Fallacy 2. The data mining process is autonomous, requiring little or no human oversight. • Reality. As we saw above, the data mining process requires significant human interactivity at each stage. Even after the model is deployed, the introduction of new data often requires an updating of the model. Continuous quality monitoring and other evaluative measures must be assessed by human analysts. Jan-Apr 2009 – w10d2

  12. Fallacies Of Data Mining • Fallacy 3. Data mining pays for itself quite quickly. • Reality. The return rates vary, depending on the startup costs, analysis personnel costs, data warehousing preparation costs, and so on. Jan-Apr 2009 – w10d2

  13. Fallacies Of Data Mining • Fallacy 4. Data mining software packages are intuitive and easy to use. • Reality. Again, ease of use varies. However, data analysts must combine subject matter knowledge with an analytical mind and a familiarity with the overall business or research model. Jan-Apr 2009 – w10d2

  14. Fallacies Of Data Mining • Fallacy 5. Data mining will identify the causes of our business or research problems. • Reality. The knowledge discovery process will help you to uncover patterns of behavior. Again, it is up to humans to identify the causes. Jan-Apr 2009 – w10d2

  15. Fallacies Of Data Mining • Fallacy 6. Data mining will clean up a messy database automatically. • Reality. Well, not automatically. As a preliminary phase in the data mining process, data preparation often deals with data that has not been examined or used in years. Therefore, organizations beginning a new data mining operation will often be confronted with the problem of data that has been lying around for years, is stale, and needs considerable updating. Jan-Apr 2009 – w10d2

  16. What Tasks Can Data Mining Accomplish? • Description • Estimation • Prediction • Classification • Clustering • Association Jan-Apr 2009 – w10d2

  17. Interestingness Measures for Data Mining: A Survey (based on Geng and Hamilton) Jan-Apr 2009 – w10d2

  18. Interestingness measures • are intended for selecting and ranking patterns • according to their potential interest to the user • regardless of the kind of patterns being mined • Interestingness is a broad concept that emphasizes • conciseness, coverage, reliability, peculiarity, diversity, novelty, surprisingness, utility, and actionability • These nine criteria can be further categorized into: • objective, subjective, and semantics-based • A concise pattern or set of patterns • is relatively easy to understand and remember and thus is added more easily to the user’s knowledge (set of beliefs) Jan-Apr 2009 – w10d2

  19. Towards comprehensive support for organizational mining (based on Song) Jan-Apr 2009 – w10d2

  20. Process mining • requires the availability of an event log • Most has been focusing on control-flow discovery • i.e., constructing a process model based on an event log • assume that it is possible to sequentially record events such that • each event refers to an activity (i.e., a well-defined step in the process) • and is related to a particular case (i.e., a process instance) • some mining techniques use additional information such as • the performer or originator of the event (i.e., the person/resource executing or initiating the activity), • the timestamp of the event, or data elements recorded with the event (e.g., the size of an order) Jan-Apr 2009 – w10d2

  21. Event logs • An interesting class of information systems • that produce event logs are • the so-called Process-Aware Information Systems (PAISs) • These systems provide very detailed information • about the activities that have been executed Jan-Apr 2009 – w10d2

  22. Organizational mining • Discovery • aims at constructing a model that reflects current situations • the organizational model • represents the current organizational structure and • the social network • shows the communication structure in an organization Jan-Apr 2009 – w10d2

  23. Organizational mining • Conformance checking • examines whether the modeled behaviour matches the observed behaviour • involves two dimension of conformance measures in the control flow perspective • fitness • is the degree of the association between the log traces and the execution paths specified by the process model • appropriateness • is the degree of accuracy with which the process • model describes observed behaviour Jan-Apr 2009 – w10d2

  24. Organizational mining • Extension • aims at enriching an existing model • by extending the model • through the projection of information • extracted from the logs onto the initial model Jan-Apr 2009 – w10d2

More Related