1 / 78

Data Mining -Introduction

This article provides an overview of data mining, including its definitions, basic functions, and its applications in association rule mining and classification. It also discusses the differences between data mining and knowledge discovery in databases (KDD).

kratochvil
Download Presentation

Data Mining -Introduction

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Mining -Introduction By Dr.M.Nandhini Assistant Professor,Dept of CS, Government Arts College,Udumalpet.

  2. Topics to be discussed What is Data Mining? ● Definitions ● Views on the Process ● Basic Functions

  3. Data Mining- Definitions - “The nontrivial extraction of implicit, previously unknown, and potentially useful information from data.” – "...the automated or convenient extraction of patterns representing knowledge implicitly stored or captured in large databases, data warehouses, the Web, ... or data streams.” – “...the process of discovering patterns in data. The process must be automatic or (more usually) semiautomatic. The patterns discovered must be meaningful...” – “...finding hidden information in a database.” – “...the process of employing one or more computer learning techniques to automatically analyze and extract knowledge from data contained within a database.”

  4. Data Mining Functions All data mining functions can be thought of as attempting to find a model to fit the data. – Predictive models predict unknown values based on known data – Descriptive modelsidentify patterns in data Each type has several sub-categories, each of which has many algorithms.

  5. Data Mining functions

  6. Data Mining vs KDD Many texts treat KDD and Data Mining as the same process, but it is also possible to think of Data Mining as the discovery part of KDD. Dunham: • KDD is the process of finding useful information and patterns in data. • Data Mining is the use of algorithms to extract information and patterns derived by the KDD process.

  7. Data Mining vs KDD

  8. What Motivated Data Mining? Huge amount of Raw DATA is available. The motivation for data mining is to • Analyse, • Classify, • Cluster, • Characterize the Data etc...

  9. Task1 :Association Rule Mining

  10. Association Rule Mining • Association rule mining is a popular and well researched method for discovering interesting relationships between item sets in large databases. • Using different measures , strong rules are discovered for finding frequent item sets. • A Database (D) consist of Transactions • Each Transactions consist of Itemsets I= (X, Y, Z …) • X Y is an Association Rule where X,Y belongs to I & X ٨ Y = Ǿ

  11. Measures: Support and Confidence • Support of X  Y: Percentage of transactions that contain X Y • Confidence of X  Y:Ratio of number of transactions that contain X  Y to the number that contain X • Large (Frequent) itemset: Itemset whose number of occurrences is above a threshold.

  12. Literature Survey Basic Methodologies Apriori Algorithm - A k-itemset is frequent only if all of its sub item sets are frequent (Downward Closure Property) Frequent Pattern tree based method - A divide-and-conquer methodology: decompose mining tasks into smaller ones - Avoid candidate generation ECLAT (Equivalence Class Transformation) - Vertical data format - Frequent Item sets are generated similar to apriori with depth first computation

  13. Apriori Algorithm Consider a particular Departmental store where it shows some transactions

  14. Apriori Algorithm The above table can be represented in binary format as below

  15. Apriori Algorithm Then the 1-item sets are generated from the binary table ..i.e.

  16. Apriori Algorithm Then by taking the support threshold value as 60% (5 transactions) i.e. minimum support count as 3, the cola and eggs are discarded from the item sets as they have less than 3(threshold). i.e. out of the 6 items in 1-item set, 2 are discarded and 4 are remaining therefore Number of frequent 2-itemsets are 4

  17. Apriori Algorithm In the 2-item sets, (beer, bread) & (beer, milk) are discarded as they have less than 3(threshold). And the 3-item sets are generated from candidate itemsets. The item set which is having more than 3 (threshold). we will not have 4-item set as there is only 1 pair formed in 3-item set. if we have 2 pairs, then the 4-item set can be formed by combining 2 pairs.

  18. Apriori Algorithm As a conclusion, Apriori generated 1,2,3-item sets as 6c1+4c2+4c3 = 6+6+4 =16 But according to brute force strategy, the same is done as 6c1+6c2+6c3=6+15+20 = 41. Hence, Apriori generates optimum and required number of item sets. Click to Exercises

  19. Task 2 :Classification

  20. CLASSIFICATION - Knowledge for Making Decisions There are three ways in which humans acquire knowledge: • learning by being told • learning by observation and discovery • learning by induction. • The knowledge engineer acquires knowledge from the domain expert by the first two means. • Knowledge acquisition is the activity whereby the machine acquires expert knowledge in order to solve some specific problems. It results in a knowledge base. • Expert systems need large chunks of knowledge to achieve high levels of performances, and the process to obtain knowledge from experts is tedious and time-consuming;

  21. The Map of Knowledge Acquisition See figure , from human know-how, there are two routes to reach the machine know-how. One is the articulation aided by knowledge engineers. The other is the rule induction performed by efficient inductive inference algorithms.

  22. A Comparison between Induction and Deduction • In logic terms, the induction task is: given Q, find a P such that P  Q • The deduction task, where given P and P  Q then know Q. • Induction is generalizing from the particular, whereas deduction is specializing from the general. • Induction is determining the formula from a set of values whereas deduction is finding a value from a formula. • Induction is the process of extracting a pattern from examples.

  23. -contd The induction task has been characterized by : • given a training set consisting of an object described by a fixed set of attributes. • given an outcome known for each object example. • find a rule that expresses the objects class in terms of the values of its attributes

  24. A specific example: • Find a means of predicting which company profiles will lead to a increase or decrease in profits based on the following data: Age, Competition and Type

  25. This problem has three attributes (Age, Competition, and Type) and one outcome/class attribute (Profit). • A simple algorithm would be to select the first attribute as the root, the next at the next level etc. till the leaf node becomes the goal. • Such a situation leads to the following rule structure: • If Age = X and Competition = Y and Type = Z then Profit = ?

  26. Classification-Definition • Given a collection of records (training set ) • Each record contains a set of attributes, one of the attributes is the class. • Find a model for class attribute as a function of the values of other attributes. • Goal: previously unseen records should be assigned a class as accurately as possible. • A test set is used to determine the accuracy of the model. Usually, the given data set is divided into training and test sets, with training set used to build the model and test set used to validate it.

  27. Classification –A Two Step Process • Model construction: Describing a set of predetermined classes. • Each tuple/sample is assumed to belong to a predefined class, as determined by the class label attribute • The set of tuples used for model construction: training set • The model is represented as classification rules, decision trees, or mathematical formulae • Model usage: for classifying future or unknown objects • Estimate accuracy of the model • The known label of test sample is compared with the classified result from the model • Accuracy rate is the percentage of test set samples that are correctly classified by the model

  28. Training Data Classifier (Model) Classification Process (1): Model Construction Classification Algorithms IF rank = ‘professor’ OR years > 6 THEN tenured = ‘yes’

  29. Classifier Testing Data Unseen Data Classification Process (2): Use the Model in Prediction (Jeff, Professor, 4) Tenured?

  30. Illustrating Classification Task

  31. Examples of Classification Task • Predicting tumor cells as malignant or not • Classifying credit card transactions as legitimate or fraudulent • Classifying secondary structures of protein as alpha-helix, beta-sheet, or random coil

  32. Methods of classification • Decision Tree based Methods • Rule-based Methods • Memory based reasoning • Neural Networks • Naïve Bayes and Bayesian Belief Networks • Support Vector Machines

  33. Decision Tree Classification Task Decision Tree

  34. categorical categorical continuous class Example of a Decision Tree Splitting Attributes Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO Model: Decision Tree Training Data

  35. NO Another Example of Decision Tree categorical categorical continuous class Single, Divorced MarSt Married NO Refund No Yes TaxInc < 80K > 80K YES NO There could be more than one tree that fits the same data!

  36. Decision Tree Classification Task Decision Tree

  37. Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO Apply Model to Test Data Test Data Start from the root of tree.

  38. Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO Apply Model to Test Data Test Data

  39. Apply Model to Test Data Test Data Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO

  40. Apply Model to Test Data Test Data Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO

  41. Apply Model to Test Data Test Data Refund Yes No NO MarSt Married Single, Divorced TaxInc NO < 80K > 80K YES NO

  42. Apply Model to Test Data Test Data Refund Yes No NO MarSt Assign Cheat to “No” Married Single, Divorced TaxInc NO < 80K > 80K YES NO

  43. Evaluating the Classifier Confusion Matrix A confusion matrix shows the number of correct and incorrect predictions made by the classification model compared to the actual outcomes (target value) in the data. The matrix is NxN, where N is the number of target values (classes).Performance of such models is commonly evaluated using the data in the matrix.

  44. Confusion Matrix Example of Confusion Matrix: 44

  45. Evaluating the Classifier • Accuracy : the proportion of the total number of predictions that were correct. Accuracy = (TP + TN)/All • Positive Predictive Value or Precision : the proportion of positive cases that were correctly identified. • Negative Predictive Value : the proportion of negative cases that were correctly identified. =TN/ (TN+FN) • Sensitivity or Recall : the proportion of actual positive cases which are correctly identified.  Sensitivity = TP/P • Specificity : the proportion of actual negative cases which are correctly identified.  Specificity = TN/N • Click for exercises

  46. ID3- Classification using Decision trees • Creates tree using information theory concepts and tries to reduce expected number of comparison.. • ID3 chooses split attribute with the highest information gain using entropy as base for calculation.

  47. Classification by Decision Trees (ID3) • The following table consists of training data from mobile store. Construct a Decision Tree based on this data, using the basic algorithm for decision tree induction. Classify the records by the “Customer satisfaction” attribute.

  48. Model Construction Iteration 1 • Root node is to identified for constructing a decision tree • Highest Information Gain attribute is selected as root node. • Overall Information Gain of the Training dataset with binary classes • Class P: customer satisfaction = “yes” • Class N: customer satisfaction = “no” • I(p, n)= • I(9,6)= -9/15 log(9/15)-6/15log(6/15) = 0.971

  49. Step-1 Attribute(A1)=“Battery Life” Entropy (A1) = 5/15 *I(5, 0)+ 5/15*I(2, 3)+ 5/15 *I(2, 5) = 5/15[ -5/5 log (5/5) -0/5 log (0/5)] + 5/15[ -2/5 log (2/5) – 3/5 log(3/5)] + 5/15[ -2/5 log (2/5) – 3/5 log(3/5)] = 0.6473 Gain(A1)= I(p, n)- E(A1)= 0.3237

  50. Step-1 Attribute(A2)= ”Memory” Entropy (A2) = 7/15 *I(3, 5)+ 8/15*I(6,2) = 0.8925 Gain(A2)= I(p, n)-E(A2)= 0.0785

More Related