1 / 32

Distributed Higher Order Association Rule Mining Using Information Extracted from Textual Data

Distributed Higher Order Association Rule Mining Using Information Extracted from Textual Data. From: Shenzhi Li, Tianhao Wu,William M. Pottenger, SIGKDD Explorations .Vol. 7, Issue 1, pp. 26-35. Presenter : Wei-Shen Tai Advisor : Professor Chung-Chian Hsu 2005/8/ 3. Outline. Introduction

enya
Download Presentation

Distributed Higher Order Association Rule Mining Using Information Extracted from Textual Data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Distributed Higher Order Association Rule Mining Using Information Extracted from Textual Data From: Shenzhi Li, Tianhao Wu,William M. Pottenger, SIGKDD Explorations.Vol. 7, Issue 1, pp. 26-35. Presenter : Wei-Shen Tai Advisor : Professor Chung-Chian Hsu 2005/8/3

  2. Outline • Introduction • Related work • D-HOTM framework • Entity extraction • Distributed High-Order ARM • A example for application • Metrics for evaluation • Conclusion and future work

  3. Information extraction and mining • Why we need them? • The burgeoning amount of textual data in distributed sources combined with the obstacles involved in creating and maintaining central repositories. • “System of Systems” • The United States Department of Homeland Security (DHS) recognizes that the proliferation of databases and schemas involving fragmented data poses a challenge to information sharing. • In response, the DHS is promulgating a approach that acknowledges the infeasibility of creating a single massive centralized database.

  4. DHS’s problems • Overlap between tiers • A three-tier structure with some local databases, state databases and federal databases. • Constraints on system interoperability and legal restrictions on data sharing, • More centralized the database, the less information that is shared. • A “System of Systems” approach • A standards for interoperability and communication in areas. • Indeed, efforts are underway to establish standards in database schema integration(e.g., OWL, GJXDM, etc.)..

  5. Mining distributed data • Current algorithms for mining distributed data • Only when the global schema across all databases is known. • Information extracted from distributed textual data • No preexisting global schema (i.e., dictionary of terms) is available. This is due to the fact that the entities extracted may differ between textual documents at the same or different locations. • Schemas of textual entities are highly fluid in nature. • As a result, a fixed global schema cannot be assumed and current algorithms that rely on the existence of such a schema cannot be employed.

  6. Distributed Association Rule Mining (ARM) • Distributed Association Rule Mining (ARM) algorithms • Mine association rules from data in horizontally or vertically fragmented databases. • Useful rules can be mined from diverse databases with different local schemas as long as records can be linked via. • Higher-order • Implies that rules may be inferred between items that do not occur in the same local database schema. • In other words, rules can be inferred based on items (entities) that may never occur together in any record in any of the distributed databases being mined.

  7. Distributed higher-order text mining(D-HOTM) • Advantage • Requires neither the knowledge of the global schema nor schema integration as a precursor to mining rules. • Extracts entities and discovers rules based on higher-order associations between entities in records linked by a common key. • Major components: • Entity extraction • Based on information extraction rules learned using a semi-supervised active learning algorithm. • Distributed association rule mining

  8. About ARM • Association rule mining (ARM) • Discovers associations between items. Given two distinct sets of items, X and Y, we say Y is associated with X if the appearance of X implies the appearance of Y in the same context. • ARM outputs a list of association rules of the format XY, where X  Y has a predetermined support and confidence. • Many ARM algorithms are based on the well-known Apriori algorithm.

  9. Parallelism in ARM (1/2) • Data parallelism • Logically partitions the database among processors. • Each processor works on its local partition using the same computational model. • Count distribution (CD) is a simple data-parallelism algorithm. Each processor generates the local candidate itemsets independently based on the local partition. • Then the global counts are computed by sharing (or broadcasting) the local counts, and the global frequent candidate itemsets are generated. • Data-parallelism exchanges only the counts among processors, which minimizes the communication cost. As a result, it seems ideal for use in a distributed environment.

  10. Parallelism in ARM (2/2) • Task parallelism • Each processor performs different computations independently, yet all have access to the entire dataset. • For example, the computation of candidate itemsets of different sizes might be distributed among processors in a parallel loop across itemsets. • In this case, each processor generates global counts independently. • In a distributed environment, this can be accomplished by performing an initial ‘copy-in’ operation of the dataset, albeit often at great cost.

  11. Fast Distributed Mining v.s. Optimized Distributed Association Mining • Fast Distributed Mining (FDM) • Based on count distribution. • The advantage of FDM over CD is that it reduces the communication cost by sending the local frequent candidate itemsets to a polling site instead of broadcasting. • Optimized Distributed Association Mining (ODAM) algorithm • Both reduces the size of the average transaction and reduces the number of message exchanges in order to achieve better performance. • The transactions are reduced by deleting the non-frequent items from the itemsets and merging several transactions with the same itemsets into one record. • As for the message exchange, instead of using broadcast as with CD or polling sites like FDM, ODAM just sends all local information to one site, called the receiver.

  12. Distributed mining of vertically fragmented data • A privacy preserving association rule mining algorithm for vertically distributed data. • The authors use a vector to represent each (vertically fragmented) record in which the attributes in the schema are distributed amongst the different local sites. • It is a method for mining association rules when the (global) schema is distributed amongst multiple sites. • This approach requires a priori knowledge of the complete (global) schema by all sites, and is thus unsuitable for mining rules from distributed data for which local schemas differ and the global schema is unknown. • Three assumptions: (1) the databases are vertically fragmented; (2) there is a global key to unambiguously link subsets of items, and no additional techniques are needed to identify which subsets should be linked; (3) the global database schema is known.

  13. Entity Extraction • reduced regular expression (RRE) • Discovers sequences of words and/or part-of-speech tags that, for a given entity, have high frequency in the labeled instances of the training data (true set) and low frequency in the unlabeled instances (false set). • First discovers the most common element1 of an RRE, termed the root of the RRE. The algorithm then extends the RRE in “AND”, “GAP”, and “Start/End” learning phases. • In each iteration, a single RRE is generated. This RRE is considered a sub-rule of the current entity.

  14. RRE Discovery Process

  15. Definition in D-HOTM • Object • In a given local database apply arow to represent it. For example a particular individual mentioned in an investigative report. • Item (or items) • Identifying the object such as a person’s name or SSN, each row also contains other items known to exist in the source document.

  16. Fragmentation types • Horizontally fragmented • There is no guarantee that every site will include the same set of items, • And in the case where an object is not a document or report but a person, different distributed sites may also refer to the same object multiple times (e.g., investigative reports about different crimes committed by the same individual). • Vertically fragmented • There is no one-to-one mapping connecting records in the distributed databases. • In addition, the (local) ‘schema’ for each individual document varies, and no clean division of all objects’ items into identical sets.

  17. Skeleton of D-HO ARM (1) Select linkage items (2) Assign a globally unique ID to each record/object (3) Identify linkable records using Apriori on global IDs (4) Exchange information about linkable records (5) For each site (6) Apply the Apriori algorithm locally

  18. Select linkage items • Items used for linking records • One requirement for this set is that the item (or combination of items) must uniquely identify objects. • In police investigative reports, SSN (suspect No.) might be selected as the subset used for linkage because such reports are written about individuals, and individuals are uniquely identified by SSN. • SSN uniquely represents person objects. • object vs. record • the items in the subset used for linkage refer to objects, while a record consists of the collection of items (i.e., entities) extracted from a given document (e.g., a police report). • Naturally this implies that one or more of the items in a given record uniquely represent the object used for record linkage.

  19. Discovers linkable records using the item(s) • For example • If a given suspect appears in a burglary record in Detroit, and the same suspect also appears in a mugging case in Philadelphia, and the SSN is chosen as the linking item, then those two distributed records are considered linkable. • Application • In a practical sense, linking these two records might reveal new information to the investigating police officer.

  20. Identifiers problem • Ideally assumption • The item (or combination of items) should be globally unique and consistent. • Unique means different objects have different identifiers while consistent implies that the records about the same object have the same identifiers. • Problems • Two databases will have the same object (a reference to a scholarly research article) with different items. • {ID, title, year, citenum} and {ID, title, year, code} • The 2-itemset {code citenum} can then be calculated using the global ID to match records in the two different databases – a higher-order association between code and citenum via the global ID.

  21. Detection of Linkable Records • If each local record corresponds to a single document object. • The combination of the title and author attributes serve both to link records from different sites and to uniquely identify objects. • Record identifiers (such as DOIs for investigative reports) differ from the object identifiers used for linking. • For example, suppose records are comprised of the entities extracted from reports that contain criminal suspect information such as SSN and modus operandi. Higher-order associations can be discovered by linking two such records through the SSN item. • In this case, the linking item, SSN, uniquely identifies individuals, not records. • A single investigative report record may contain multiple SSNs.

  22. Theory for the Detection of Linkable Records (1/3) • Definition 1: Given two records Ri and Rj, let T= Ri Rj, and T. For any item aT, we say Ri and Rj are 2nd-order linkable through a, or a2-linkable, denoted as Ri ~a Rj. • Theorem 1: The relation between records that are a2-linkable is reflexive, symmetric and transitive.

  23. Theory for the Detection of Linkable Records (2/3) • Definition 2: Given n distinct records R1, R2, …, Ri, …, Rn and Ti= Ri Ri+1 with Ti, let L be a list of items L = (a1, a2, …, ai,…, an-1) such that aiTi. Then we say records R1 and Rn are (k+1)th-order linkable through L, where k is the number of distinct items in L. The (k+1)th-order link is denoted R1 ~a1 R2 ~a2 R3 ~a3 R4 … Ri-1 ~a(i-1) Ri … Rn-1 ~a(n-1) Rn, or in short, R1 ~L Rn. We term L a viable path. • Theorem 2: For any higher-order link between two records, there must exist at least one link which does not have repeated items in the viable path, or repeated records in the link.

  24. Theory for the Detection of Linkable Records (3/3) • Definition 3: The length of a higher-order link cluster is defined as the number of groups in the cluster. The distance between two groups is defined as the shortest of all possible higher-order links. • For example, given the following two higher-order link clusters: • Ra ~a Rab ~b Rbc ~c Rc • Ra ~a Rac ~c Rc • The length of the first cluster is four while the length of the second is three. Suppose these two link clusters contain the only links between Ra and Rc, then from the link cluster Ra ~a Rac ~c Rc. we see that the distance between Ra and Rc is three. • Theorem 4: Given a frequent k-itemset X generated by Apriori for k2, for any pair of items which are members of X, the distance between the groups on those items is at most three. • Theorem 5: Given a frequent k-itemset A and j-itemset B (j,k2) for which A is not a subset of B and B is not a subset of A, suppose there exists at least one item which is a member of A and B. Furthermore, suppose there are items aA and bB, then the distance between groups Ra and Rb is at most four.

  25. An Example Application

  26. Link results • The distance between RJack and RDiana is three, and the groups are higher-order linked as follows: RJack-Jack,Diana ~Jack RJack,Diana ~Diana RDiana-Jack,Diana. • Using the GIDLists to represent the groups on the items, we have • {2,6,8}-{6} ~Jack {6} ~Diana {4,6}-{6}, or {2,8}~Jack {6} ~Diana {4}.

  27. Merged records based on the higher-order links discovered • The Apriori algorithm is applied again, this time to each local database. Since higher-order associations have been implicitly included in the new, merged records, both higher-order and the usual first-order associations will be included in the resulting rules generated by Apriori. • For example, “Gun Þ Robbery” and “Diana Þ Robbery” are rules generated based on higher-order associations discovered by the D-HO ARM algorithm.

  28. Metrics for evaluation • Support value • One of the most important metrics used in ARM, which is defined as the frequency of an itemset divided by the total number of instances. • In distributed databases, the total number of instances can be calculated by counting the number of unique global object IDs. • Problems • Different objects (and subsequently records) can be incorrectly mapped to the same global ID • Records that should be mapped to a single global ID can be mapped to different IDs. • The error rate of the mapping function will thus influence both the support and confidence metrics.

  29. An example of mapping • The ground truth data set is obtained through experts’ identification. • The modeled data set is obtained through the mapping function. • The sample data set is R={1,2,3,4,5,6}, where records 1 and 2 represent the same object (e.g., have the same primary key), records 3 and 4 represent another object, and records 5 and 6 represent a third object.

  30. Error rate and difference degree • Observed error rate f is: • We also define the observed difference degree t’ as: • where n’ is the size of the modeled data set and n is the size of the ground truth.

  31. Conclusions • Heterogeneous distributed databases can be easily centralized. • Distributed higher-order rules can be mined from multiple diverse databases without knowledge of the global schema. • Propose a support metric in the context of distributed higher-order ARM. • Mining higher-order rules from distributed textual data in a hybrid fragmented form independent of the knowledge of the global schema.

  32. Future work • ID identification automatically • This process could be partially automated given semantic mappings such as those supported by the recently released Web Ontology Language. • Distributed ARM deals with the efficiency of computation • Make workload balancing as well as optimization of both computation and communication. • Opposite direction mining • Be used to mine association rules from the results returned from a search (as opposed to mining rules from entire databases). • For example, a user may wish to find strong associations between Axiom’s classification codes and the number of citations on Citeseer for a particular set of documents returned by queries. • Multilevel parallelism • Can be introduced to improve runtime performance: e.g., at each distributed database, a parallel ARM algorithm can be executed.

More Related