1 / 34

CSR : Discovering Subsumption Relations for the Alignment of Ontologies

CSR : Discovering Subsumption Relations for the Alignment of Ontologies. Vassilis Spiliopoulos 1, 2 , Alexandros G. Valarakos 1 , and George A. Vouros 1 1 AI Lab Department of Information and Communication Systems Eng. University of the Aegean 83200 Karlovassi, Samos, Greece

daphne
Download Presentation

CSR : Discovering Subsumption Relations for the Alignment of Ontologies

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CSR: Discovering Subsumption Relations for the Alignment of Ontologies Vassilis Spiliopoulos1, 2, Alexandros G. Valarakos1, and George A. Vouros1 1 AI Lab Department of Information and Communication Systems Eng. University of the Aegean 83200 Karlovassi, Samos, Greece {vspiliop, georgev}@aegean.gr

  2. Outline • Introduction • Problem Definition • Why Subsumption Relations • Related Work • The Method • Experimental Results • Conclusions

  3. Ontology • Concept features • Properties • Data type • Object Property (relation) • Instances • Comments • Conceptsorganized into hierarchies (subsumption relation) • Ontology Languages • OWL Family • Union, Intersection, Disjointness date of Publication Reference “⊑” “⊑” # of pages title title Proceedings Book The Semantic Web 08 Proc. 973 “⊑” “⊑” Edited Selection Monograph A book that is collection of texts or articles

  4. Current Situation ... ... Engineer 1 Engineer 2 Engineer N Bibliographic Domain ... ontology 1 ontology N

  5. Ontology Mapping “Find Citations in Proceedings” • Ontology Mapping is a process that has as input two ontologies and locates relations (i.e. mappings) between their elements • Equivalence (≡) • Subsumption (⊑ or ⊒) • Intersection (⊥) Locates a mapping Retrieves a superset of what he is looking for Conference Ontology Agents’ Ontology date of date Publication Reference Work Citation to # of pages title # of pages title Proceedings Book title title Book Proceedings “⊑”

  6. Why Subsumption Relations (1/2) “≡” “≡” Work date of Citation date to Publication Reference “⊑” “⊑” # of pages title title “⊑” Book title # of pages title Proceedings Proceedings Book “⊥” “⊑” Monograph Edited Selection Edited Selection Monograph chapters chapters “≡” # of pages

  7. Why Subsumption Relations (2/2) • Discover subsumption relations separately from subsumptions and equivalencies that can be deduced by a reasoning mechanism • May augment the effectiveness of current ontology mapping and merging methods • No or few equivalences • Web Service matchmaking • Ontology engineering environments

  8. Problem Definition • The subsumption computation problem is defined as follows: • Given two input ontologies • optionally,specifying properties’ equivalences • Classify each pair (C1,C2) of concepts to two distinct classes: To the “subsumption” (⊑) class (C1 ⊑ C2 ), or to the class “R” • Class “R” denotes pairs of concepts that are not known to be related via the subsumption relation

  9. Related Work • Satisfiability Based Approaches [1] • Transformation of the ontology mapping problem in a satisfiability one • Exploitation of Domain Knowledge [2], [3], and [4] • Exploit domain ontologies as an intermediate ontology for bridging the semantic gap [2], [3] • WordNet is used for the same purpose (WordNet Description Logics) [4] • Google Based Approaches [5], [6], and [7] • Exploit the hits returned by Google to test if subsumption relation holds [5], [6] or to loosen the formal constrains [7] • Machine Learning Approaches [8] • A method based on Implication Intensity theory (Unsupervised Learning) is proposed

  10. The CSR Method At a Glance (1/2) • Purpose • We try to learn patterns of features that indicate a subsumption relation between two concepts belonging to two different ontologies • How • By exploiting supervised machine learning techniques (binary classification), • and the ontology specification semantics

  11. The CSR Method At a Glance (2/2) • Why machine learning? • There are no evident generic rules directly capturing the existence of a subsumption relation (e.g. labels/vicinity similarity or dissimilarity) • Learn patterns of features not evident to the naked eye • Self-adapting to idiosyncrasies of specific domains • Non-dependant to external resources

  12. Generation of Testing Pairs (Search Space Pruning) Hierarchies Enhancement Generation of Features SEMA Generation of Training Examples Train Classifier The CSR Method (1/12) • Input • Two OWL-DL ontologies (the process is not language specific) • Optionally, property equivalencies computed by SEMA mapping tool • The method requires the existence of subsumption relations between concepts R

  13. Generation of Testing Pairs (Search Space Pruning) Generation of Features SEMA Generation of Training Examples Train Classifier The CSR Method (2/12) • Hierarchies Enhancement • Inferring all indirect subsumption relations • Influences the generation of training examples and feature vectors Hierarchies Enhancement R

  14. Generation of Testing Pairs (Search Space Pruning) SEMA Generation of Training Examples Train Classifier The CSR Method (3/12) • Generation of Features • CSR exploits two types of features: Concepts’ properties or words appearing in the “vicinity” of concepts Hierarchies Enhancement R Generation of Features

  15. The CSR Method (4/12) p1 p2 pN O1 p2 p1 (f1, f2, ..., fN) fi: i-th feature pi O2 pN pN-1 (C1, C2)

  16. The CSR Method (5/12) w1 w2 wT (frj1, frj2, ... , frjT) O1 w1 w3 (2, 0, 1, ... , 0) C1 w1 ... O2 C2 ... w wT (0, 0, 1, ... , 2) w3 wT • For each concept • Label • Comments • Properties • Instances • Related Concepts ... ... wT CM wi (0, 1, 0, ...,1 , … , 1) ... w2 fri: frequency of i-th word T: number of distinct words

  17. The CSR Method (6/12) Left Side Concept Right Side Concept C1 C2 (fr11, fr12, ..., fr1i , ... , fr1T) (fr21, fr22, ..., fr2i, ... , fr2T) (C1, C2) (f1, f2, ..., fi ,..., fT)

  18. Generation of Testing Pairs (Search Space Pruning) SEMA Generation of Training Examples Train Classifier The CSR Method (7/12) • Generation of Training Examples • Classes: “⊑” and R • Training examples are being generated by exploiting the input ontologies in isolation • According to the semantics of specifications Hierarchies Enhancement R Generation of Features

  19. The CSR Method (8/12) • Class “⊑” • Subsumption Relation. Include all concept pairs from both input ontologies that belong in the subsumption relation (direct or indirect) • Equivalence Relation. Any concept in a training pair can be substituted by any of its equals • Union Constructor. E.g. C4 ⊔ C5 ⊑ C2 => C4⊑C2 and C5⊑C2

  20. The CSR Method (9/12) • Generic class “R” • If there is not an axiom that specifies the subsumption relation between a pair of concepts • Categories of class “R” • Concepts belonging to different hierarchies • Siblings at the same hierarchy level • Siblings at different hierarchy levels • Concepts related through a non-subsumption relation • Inverse pairs of class “⊑”

  21. The CSR Method (10/12) • Balancing the Training Dataset • The number of training examples for the class “⊑” are much less than the ones for class “R” • Dataset imbalance problem • Two balancing strategies: • Random under-sampling variation • Random over-sampling

  22. Generation of Testing Pairs (Search Space Pruning) SEMA Generation of Training Examples Train Classifier The CSR Method (11/12) Hierarchies Enhancement R Generation of Features

  23. Generation of Testing Pairs (Search Space Pruning) SEMA Generation of Training Examples Train Classifier The CSR Method (12/12) Hierarchies Enhancement R Generation of Features 1st Ontology 2nd Ontology C11 C12 C13 C28 C21 C22 C29 C211 C210 C23 C24 C26 C27 C25

  24. Experimental Settings • The testing dataset has been derived from the benchmarking series of the OAEI 2006 contest • The compiled corpus + gold standard is available at http://www.icsd.aegean.gr/incosys/csr • Classifiers used: C4.5, Knn (2 neighbors), NaiveBayes (Nb) and Svm (radial basis kernel) • We denote each type of experiment withA+B+C • A: classifier, • B: type of features(“Props” for properties or “Terms” for words) and • C: dataset balancing method (“over” and “under” for over- and under-sampling) • Baseline: Consults the vectors of the training examples of the class “⊑”, and selects the first exact match (No generalization) • Description Logics’ Reasoner: We specify axioms concerning only properties’ equivalencies (Reasoner+Props), or alternatively, both properties’ and concepts’ equivalencies (Reasoner+Props+Con)

  25. Overall Results • All classifiers (except Svm) based on properties outperform Baseline+Props • Generalization – location of pairs not in the training dataset

  26. Overall Results • All classifiers based on words outperform Baseline+Words • Generalization – location of pairs not in the training dataset

  27. Overall Results • C4.5 (using both words or properties) performs best comparing to all other CSR experimentation settings • Disjunctive descriptions of cases: More than one features may indicate whether a specific concept pair belongs in the class “⊑” • Decision trees are very tolerant to errors in the training set. Both to feature vectors and training examples

  28. Overall Results • CSR exploiting words does not require neither properties nor concepts equivalencies • Reasoner exploits such equivalencies • Depends on the mapping tool

  29. Closer Look • A7 Category • Different conceptualizations • Flattened classes in target ontology + props defined in a more detailed manner • SEMA: 74% Precision – 100% Recall (Props+Cons)

  30. Closer Look • R1-R2 Category • Different conceptualizations • CSR locates subsumptions that the reasoner cannot infer (R2), without using equivalencies

  31. “Confused” Equivalencies • CSR is very tolerant in “confusing” equivalence relations as subsumption ones • Without using them also as input • Can be used for filtering

  32. Conclusions • CSR method: • Learns patterns of concepts’ features (properties or terms) that provide evidence for the subsumption relation among these concepts, using machine learning techniques • Generates training datasets from the source ontologies specifications • Tackles the problem of imbalanced training datasets • Generalizes effectively over the training examples • Does not exploits equivalence mapping(words case as features) • Does not easily “confuse” equivalence mappings as subsumption ones • Is independent of external resources

  33. Questions? Comments? Thank you!

  34. Related Work • Giunchiglia, F., Yatskevich, M., Shvaiko, P.: Semantic Matching: Algorithms and implementation. Journal on Data Semantics, IX (2007) • Aleksovski, Z., Klein, M., Kate, W, Harmelen F.: Matching Unstructured Vocabularies Using a Background Ontology. In: EKAW, Podebrady, Czech Republic (2006) • Gracia, J., Lopez, V., D'Aquin, M., Sabou, M, Motta, E., Mena, E.: Solving Semantic Ambiguity to Improve Semantic Web based Ontology Matching, In: Ontology Matching Workshop, Busan, Korea (2007) • Bouquet, P., Serafini, L., Zanobini, S., and Sceffer, S. 2006: Bootstrapping semantics on the web: meaning elicitation from schemas. In: WWW, Edinburgh, Scotland (2006) • Cimiano P., Staab, S.: Learning by googling, In: SIGKDD Explor. Newsl., USA (2004) • Hage, W.R. Van, Katrenko, S., Schreiber, A.Th.: A Method to Combine Linguistic Ontology Mapping Techniques, In: ISWC, Osaka, Japan (2005) • Risto G., Zharko A., Warner K.: Using Google Distance to weight approximate ontology matches. In: WWW, Banff, Alberta, Canada (2007) • Jerome D., Fabrice G., Regis G., Henri B.: An interactive, asymmetric and extensional method for matching conceptual hierarchies. In: EMOI – INTEROP Workshop, Luxem-bourg (2006)

More Related