1 / 17

Multifaceted Exploitation of Metadata for Attribute Match Discovery in Information Integration

Multifaceted Exploitation of Metadata for Attribute Match Discovery in Information Integration. David W. Embley David Jackman Li Xu. Background . Problem : Attribute Matching Matching Possibilities (Facets) Attribute Names Data-Value Characteristics Expected Data Values

boyer
Download Presentation

Multifaceted Exploitation of Metadata for Attribute Match Discovery in Information Integration

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Multifaceted Exploitation of Metadata for Attribute Match Discovery in Information Integration David W. Embley David Jackman Li Xu

  2. Background • Problem : Attribute Matching • Matching Possibilities (Facets) • Attribute Names • Data-Value Characteristics • Expected Data Values • Data-Dictionary Information • Structural Properties

  3. Approach • Target Schema T • Source Schema S • Framework • Individual Facet Matching • Combining Facets • Best-First Match Iteration

  4. Year Year Year Year Make Make Make Feature Make has has has has has 0:1 0:1 0:1 0:1 0:* 0:1 0:1 Car Cost Model Model Model Car Model has has 0:1 has 0:1 has Phone Mileage Miles Example Car Car Style 0:1 has 0:* 0:1 0:1 has has has Mileage Miles Cost Target Schema T Source Schema S

  5. Individual Facet Matching • Attribute Names • Data-Value Characteristics • Expected Data Values

  6. Attribute Names • Target and Source Attributes • T : A • S : B • WordNet • C4.5 Decision Tree: feature selection • f0: same word • f1: synonym • f2: sum of distances to a common hypernym root • f3: number of different common hypernym roots • f4: sum of the number of senses of A and B

  7. The number of different common hypernym roots of A and B The sum of the number of senses of A and B The sum of distances of A and B to a common hypernym WordNet Rule

  8. Confidence Measures

  9. Data-Value Characteristics • C4.5 Decision Tree • Features • Numeric data (Mean, variation, standard deviation, …) • Alphanumeric data (String length, numeric ratio, space ratio)

  10. Confidence Measures

  11. Expected Data Values • Target Schema T and Source Schema S • Regular expression recognizer for attribute A in T • Data instances for attribute B in S • Hit Ratio = N’/N for (A, B) match • N’ : number of B data instances recognized by the regular expressions of A • N: number of B data instances

  12. Confidence Measures

  13. 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 Combined Measures Threshold: 0.5

  14. Final Confidence Measures

  15. F1 93.75% F2 84% F3 92% F1 98.9% F2 97.9% F3 98.4% Experimental Results • Matched Attributes • 100% (32 of 32); • Unmatched Attributes • 99.5% (374 of 376); • “Feature” ---”Color”; • “Feature” ---”Body Type”.

  16. Conclusions • Direct Attribute Matching – feasible • Individual-Facet Matching – good • Multifaceted Matching – better

  17. Future Work • Additional Facets • More Sophisticated Combinations • Additional Application Domains • Automating Feature Selection • Indirect Attribute Matching www.deg.byu.edu

More Related