1 / 33

Classifier performance & Model comparison Prof. Navneet Goyal BITS, Pilani

Classifier performance & Model comparison Prof. Navneet Goyal BITS, Pilani. Classifier Accuracy. Measuring Performance Classification accuracy on test data Confusion matrix OC Curve Classifier Accuracy Measures Evaluating Accuracy Ensemble Methods – Increasing Accuracy Boosting

Download Presentation

Classifier performance & Model comparison Prof. Navneet Goyal BITS, Pilani

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Classifier performance & Model comparisonProf. NavneetGoyalBITS, Pilani

  2. Classifier Accuracy • Measuring Performance • Classification accuracy on test data • Confusion matrix • OC Curve • Classifier Accuracy • Measures • Evaluating Accuracy • Ensemble Methods – Increasing Accuracy • Boosting • Bagging

  3. Measuring Performance

  4. Classification Performance True Positive True Negative False Positive False Negative

  5. Model Evaluation • Metrics for Performance Evaluation • How to evaluate the performance of a model? • Methods for Performance Evaluation • How to obtain reliable estimates? • Methods for Model Comparison • How to compare the relative performance among competing models?

  6. Operating Characteristic Curve • OC Curve (operating characteristic curve) or ROC curve (receiver operating characteristic curve) • Developed in 1950s for signal detection theory to analyze noisy signals • Characterize the trade-off between positive hits and false alarms • Shows relationship between FPs and TPs • Originally used in communications area to examine false alarm rates • Used in IR to examine fallout (%age of retrieved that are not relevant) versus recall (%age of retrieved that are relevant)

  7. ROC (Receiver Operating Characteristic) • Developed in 1950s for signal detection theory to analyze noisy signals • Characterize the trade-off between positive hits and false alarms • ROC curve plots TP (on the y-axis) against FP (on the x-axis) • Performance of each classifier represented as a point on the ROC curve • changing the threshold of algorithm, sample distribution or cost matrix changes the location of the point

  8. At threshold t: TP=0.5, FN=0.5, FP=0.12, FN=0.88 ROC Curve - 1-dimensional data set containing 2 classes (positive and negative) - any points located at x > t is classified as positive

  9. ROC Curve (TP,FP): • (0,0): declare everything to be negative class • (1,1): declare everything to be positive class • (1,0): ideal • Diagonal line: • Random guessing • Below diagonal line: • prediction is opposite of the true class

  10. Using ROC for Model Comparison • No model consistently outperform the other • M1 is better for small FPR • M2 is better for large FPR • Area Under the ROC curve • Ideal: • Area = 1 • Random guess: • Area = 0.5

  11. How to Construct an ROC curve? • Use classifier that produces posterior probability for each test instance P(+|A) • Sort the instances according to P(+|A) in decreasing order • Apply threshold at each unique value of P(+|A) • Count the number of TP, FP, TN, FN at each threshold • TP rate, TPR = TP/(TP+FN) • FP rate, FPR = FP/(FP + TN)

  12. How to Construct an ROC curve? Threshold >= ROC Curve:

  13. Test of Significance • Given two models: • Model M1: accuracy = 85%, tested on 30 instances • Model M2: accuracy = 75%, tested on 5000 instances • Can we say M1 is better than M2? • How much confidence can we place on accuracy of M1 and M2? • Can the difference in performance measure be explained as a result of random fluctuations in the test set?

  14. Operating Characteristic Curve

  15. Confusion Matrix Example Using height data example with Output1 correct and Output2 actual assignment

  16. Accuracy of a classifier on a given test set is the %age of test set tuples that are correctly classified by the classifier • In pattern recognition literature, it is called ‘recognition rate’ • Reflects how well the classifier recognizes tuples of various classes • Error rate or misclassification rate of a classifier M is 1 – Acc(M), where Acc(M) is the accuracy of M Classifier Accuracy Measures

  17. Example: we trained a classifier to classify medical data tuples as either ‘cancer’ or ‘not_ cancer’ • 90% accuracy rate – sounds good • What if only 3-4% of traning set tuples are actually ‘cancer’? • Classifier could be correctly labeling only the not_cancer tuples • How well the classifier recognizes ‘cancer’ tuples (+ tuples) and how well it can recognize not_cancer tuples (- tuples)? • USE Sensitivity & Specificity respectively Alternative Measures

  18. Alternative Measures • Sensitivity (True + or recognition) • Specificity (False +) • In addition, we may use ‘precision’ to assess the %age of tuples labeled as ‘cancer’ that actually are cancer tuples • Sensitivity = #(True +)/+ • Specificity = #(True -)/- • Precision = #(True +)/(#(True +) + #(False +)) • where #(True +) is the no. of true positives (cancer tuples that were correctly classified) • + is the no. of ‘cancer’ tuples • #(True -) is the no. of true negatives (not_cancer tuples that were correctly classified as such) • - is the no. of ‘not_cancer’ tuples • #(False +) is the no. of false positives (‘not_cancer’ incorrectly labelled as ‘cancer’)

  19. It can be shown that Accuracy os a function of Sensitivity and Specificity • Accuracy = Sensitivity ((pos)/(pos+neg) ) • + Specoficity ((neg)/(pos+neg) ) • Other cases where ‘accuracy’ not appropriate? • Assumption: each tuples belongs to only one class? • Rather than returning a class label, it is useful to return a probability class distribution • A classification is considered correct if it assigns a tuple to first or the second most probable class Alternative Measures

  20. Data sets with skewed class distributions are common in real life applications • Defective products in an assembly lines • Fraudulent credit card transactions are outnumbered by legitimate transactions • Malicious attacks on web sites • In both cases, there are disproportionate no. of instances that belong to different classes • Six-sigma principle: 4 defects in 1m products • 1% fraudulent transactions • Imbalance causes a no. of problems to existing classification algorithms • Accuracy measures not appropriate!! Class Imbalance Problem

  21. Metrics for Performance Evaluation • Focus on the predictive capability of a model • Rather than how fast it takes to classify or build models, scalability, etc. • Confusion Matrix: a: TP (true positive) b: FN (false negative) c: FP (false positive) d: TN (true negative)

  22. Metrics for Performance Evaluation… • Most widely-used metric:

  23. Limitation of Accuracy • Consider a 2-class problem • Number of Class 0 examples = 9990 • Number of Class 1 examples = 10 • If model predicts everything to be class 0, accuracy is 9990/10000 = 99.9 % • Accuracy is misleading because model does not detect any class 1 example

  24. Cost Matrix C(i|j): Cost of misclassifying class j example as class i

  25. Computing Cost of Classification Accuracy = 80% Cost = 150x-1+60x1+40x100 =3910 Accuracy = 90% Cost = 250x-1+5x1+45x100 =4255

  26. Accuracy is proportional to cost if1. C(Yes|No)=C(No|Yes) = q 2. C(Yes|Yes)=C(No|No) = p N = a + b + c + d Accuracy = (a + d)/N Cost = p (a + d) + q (b + c) = p (a + d) + q (N – a – d) = q N – (q – p)(a + d) = N [q – (q-p)  Accuracy] Cost vs Accuracy

  27. Cost-Sensitive Measures • Precision is biased towards C(Yes|Yes) & C(Yes|No) • Recall is biased towards C(Yes|Yes) & C(No|Yes) • F-measure is biased towards all except C(No|No)

  28. Holdout • Random Sub-Sampling • Cross Validation • Bootstrap Evaluating Classifier Performance

  29. Given data randomly partitioned in to a training set and a test set • 2/3 training set • 1/3 test set • Training set is used to derive the model • Accuracy is tested using the test set • Estimate is pessimistic since only a part of the given data is used for building the model • Sensitive to the choice of training and test sets • Confidence interval is large for varying sizes of training and test sets • Over representation in one leads to under representation in the other & vice versa Holdout Method

  30. Variation of the holdout method • Holdout method is repeated k times • Overall accuracy is the average of the accuracies from each iteration • Less sensitive to the choice of training and test data sets unlike the hold out method • No control over the no. of times each record is used for training and testing Random Subsampling

  31. k-fold cross validation • Randomly partition data into k mutually exclusive subsets or ‘folds’ D1, D2, …, Dk • Approximately equal size • Training and testing is performed k times • In iteration i, the partition Di is reserved for testing, and the remaining k-1 subsets are used to train the model • Unlike holdout and random sub-sampling methods, each sample is used the same number of times for training and once for testing • Accuracy = overall no. of correct classifications/total no. of tuples in initial data • LEAVE ONE OUT (k=N) – each test set contains only 1 record. • Test set mutually exclusive and effectively cover entire data set • Computationally expensive Cross-validation

  32. Samples the given training tuples uniformly with replacement • Each time a tuple is selected , it is equally likely be selected again and readded to the training set • You are allowed to select the same tuple more than once • Several bootstrap methods • Most common is .632 bootstrap • Data set of d tuples • The Set is sampled d times, with replacement, resulting in a bootstrap sample or training set of d samples Bootstrap

  33. Likely that some of the original data tuples will occur more than once in this sample • The tuples that did not make it to the training set form the test set • Try this out several times • On an average, 63.2% of tuples will end up in the bootstrap, and the remaining 36.8% will form the test set • Where from this 0.632 come???? • Prob. that a record is chosen by bootstrap is • 1-(1-1/N)N when N∞ is 1-1/e = 0.632 Bootstrap

More Related