1 / 46

Applications of Data Mining in Microarray Data Analysis

Applications of Data Mining in Microarray Data Analysis. Yen-Jen Oyang Dept. of Computer Science and Information Engineering. Observations and Challenges in the Information Age. A huge volume of information has been and is being digitized and stored in the computer.

Download Presentation

Applications of Data Mining in Microarray Data Analysis

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Applications of Data Mining in Microarray Data Analysis Yen-Jen Oyang Dept. of Computer Science and Information Engineering

  2. Observations and Challenges in the Information Age • A huge volume of information has been and is being digitized and stored in the computer. • Due to the volume of digitized information, effectively exploitation of information is beyond the capability of human being without the aid of intelligent computer software.

  3. An Example of Data Mining • Given the data set shown on next slide, can we figure out a set of rules that predict the classes of objects?

  4. Data Set

  5. × × × × × 。 。 。 × × 。 。 。 。 × 。 。 。 30 × × × × × × 10 15 20 Distribution of the Data Set

  6. Rule Based on Observation

  7. Rule Generated by a RBF(Radial Basis Function) Network Based Learning Algorithm Let and If then prediction=“O”. Otherwise prediction=“X”.

  8. Identifying Boundary of Different Classes of Objects

  9. Boundary Identified

  10. Data Mining /Knowledge Discovery • The main theme of data mining is to discover unknown and implicit knowledge in a large dataset. • There are three main categories of data mining algorithms: • Classification; • Clustering; • Mining association rule/correlation analysis.

  11. Data Classification • In a data classification problem, each object is described by a set of attribute values and each object belongs to one of the predefined classes. • The goal is to derive a set of rules that predicts which class a new object should belong to, based on a given set of training samples. Data classification is also called supervised learning.

  12. Instance-Based Learning • In instance-based learning, we take k nearest training samples of a new instance (v1, v2, …, vm) and assign the new instance to the class that has most instances in the k nearest training samples. • Classifiers that adopt instance-based learning are commonly called the KNN classifiers.

  13. Example of the KNN • If an 1NN classifier is employed, then the prediction of “” = “X”. • If an 3NN classifier is employed, then prediction of “” = “O”.

  14. Applications of Data Classification in Bioinformatics • In microarray data analysis, data classification is employed to predict the class of a new sample based on the existing samples with known class.

  15. For example, in the Leukemia data set, there are 72 samples and 7129 genes. • 25 Acute Myeloid Leukemia(AML) samples. • 38 B-cell Acute Lymphoblastic Leukemia samples. • 9 T-cell Acute Lymphoblastic Leukemia samples.

  16. Gene1 Gene2 ‧‧‧‧‧‧ Genen Sample1Sample2 Samplem Model of Microarray Data Sets

  17. Alternative Data Classification Algorithms • Decision tree (Q4.5 and Q5.0); • Instance-based learning(KNN); • Naïve Bayesian classifier; • Support vector machine(SVM); • Novel approaches including the RBF network based classifier that we have recently proposed.

  18. Accuracy of Different Classification Algorithms

  19. Comparison of Execution Time(in seconds)

  20. More Insights

  21. Data Clustering • Data clustering concerns how to group a set of objects based on their similarity of attributes and/or their proximity in the vector space. Data clustering is also called unsupervised learning.

  22. The Agglomerative Hierarchical Clustering Algorithms • The agglomerative hierarchical clustering algorithms operate by maintaining a sorted list of inter-cluster distances. • Initially, each data instance forms a cluster. • The clustering algorithm repetitively merges the two clusters with the minimum inter-cluster distance.

  23. Upon merging two clusters, the clustering algorithm computes the distances between the newly-formed cluster and the remaining clusters and maintains the sorted list of inter-cluster distances accordingly. • There are a number of ways to define the inter-cluster distance: • minimum distance (single-link); • maximum distance (complete-link); • average distance; • mean distance.

  24. An Example of the Agglomerative Hierarchical Clustering Algorithm • For the following data set, we will get different clustering results with the single-link and complete-link algorithms. 5 1 3 4 2 6

  25. Result of the Single-Link algorithm 5 1 3 4 2 6 3 1 4 5 2 6 Result of the Complete-Link algorithm 5 1 3 4 2 6 3 1 2 4 5 6

  26. Remarks • The single-link and complete-link are the two most commonly used alternatives. • The single-link suffers the so-called chaining effect. • On the other hand, the complete-link also fails in some cases.

  27. Example of the Chaining Effect Single-link (10 clusters) Complete-link (2 clusters)

  28. Effect of Bias towards Spherical Clusters Single-link (2 clusters) Complete-link (2 clusters)

  29. K-Means: A Partitional Data Clustering Algorithm • The k-means algorithm is probably the most commonly used partitional clustering algorithm. • The k-means algorithm begins with selecting k data instances as the means or centers of k clusters.

  30. The k-means algorithm then executes the following loop iteratively until the convergence criterion is met. • repeat { • assign every data instance to the closest cluster based on the distance between the data instance and the center of the cluster; • compute the new centers of the k clusters; • } until(the convergence criterion is met);

  31. A commonly-used convergence criterion is

  32. initial center initial center initial center Illustration of the K-Means Algorithm---(I)

  33. x new center after 1st iteration x new center after 1st iteration x new center after 1st iteration Illustration of the K-Means Algorithm---(II)

  34. new center after 2nd iteration new center after 2nd iteration new center after 2nd iteration Illustration of the K-Means Algorithm---(III)

  35. A Case in which the K-Means Algorithm Fails • The K-means algorithm may converge to a local optimal state as the following example demonstrates: Initial Selection

  36. Remarks • As the examples demonstrate, no clustering algorithm is definitely superior to other clustering algorithms with respect to clustering quality.

  37. Applications of Data Clustering in Microarray Data Analysis • Data clustering has been employed in microarray data analysis for • identifying the genes with similar expressions; • identifying the subtypes of samples.

  38. Feature Selection in Microarray Data Analysis • In microarray data analysis, it is highly desirable to identify those genes that are correlated to the classes of samples. • For example, in the Leukemia data set, there are 7129 genes. We want to identify those genes that lead to different disease types.

  39. Furthermore, Inclusion of features that are not correlated to the classification decision may result in lower classification accuracy or poor clustering quality. • For example, in the data set shown on the following page, inclusion of the feature corresponding to the Y-axis causes incorrect prediction of the test instance marked by “”, if a 3NN classifier is employed.

  40. y • It is apparent that “o”s and “x” s are separated by x=10. If only the attribute corresponding to the x-axis was selected, then the 3NN classifier would predict the class of “” correctly. x x=10

  41. Univariate Analysis in Feature Selection • In the univariate analysis, the importance of each feature is determined by how objects of different classes are distributed in this particular axis. • Let and denote the feature values of class-1 and class-2 objects, respectively. • Assume that the feature values of both classes of objects follow the normal distribution.

  42. Then, is a t-distribution with degree of freedom = (m+n-2), where If the t statistic of a feature is lower than a threshold, then the feature is deleted.

  43. Multivariate Analysis • The univariate analysis is not able to identify crucial features in the following example.

  44. Therefore, multivariate analysis has been developed. However, most multivariate analysis algorithms that have been proposed suffer high time complexity and may not be applicable in real-world problems.

  45. Summary • Data clustering and data classification have been widely used in microarray data analysis. • Feature selection is the most challenging issue as of today.

More Related