1 / 56

UMass and Learning for CALO

UMass and Learning for CALO. Andrew McCallum Information Extraction & Synthesis Laboratory Department of Computer Science University of Massachusetts. Outline. CC-Prediction Learning in the wild from user email usage DEX

mizell
Download Presentation

UMass and Learning for CALO

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. UMass andLearning for CALO Andrew McCallum Information Extraction & Synthesis Laboratory Department of Computer Science University of Massachusetts

  2. Outline • CC-Prediction • Learning in the wild from user email usage • DEX • Learning in the wild from user correction...as well as KB records filled by other CALO components • Rexa • Learning in the wild from user corrections to coreference... propagating constraints in a Markov-Logic-like system that scales to ~20 million objects • Several new topic models • Discover interesting useful structure without the need for supervision... learning from newly arrived data on the fly

  3. CC Prediction Using Various Exponential Family Factor Graphs Learning to keep an org. connected & avoid stove-piping. First steps toward ad-hoc team creation. Learning in the wild from user’s CC behavior,and from other parts of the CALO ontology.

  4. Graphical Models for Email • Compute P(y|x) for CC prediction - function - random variable - N replications Recipient of Email y N The graph describes the joint distribution of random variables in term of the product of local functions xb xs xr Nb Ns Nr-1 Email Model:Nb words in the body, Ns words in the subject, Nr recipients Body Subject Other Words Words Recipients Nr • Local functions facilitate system engineering through modularity

  5. Document Models • Models may relational attributes Na Author ofDocument y xb xs xb xr xt Nb Ns Na-1 Nr Nt Title Abstract Body Co-authors References • We can optimize P(y|x) for classification performance and P(x|y) for model interpretability and parameter transfer (to other models)

  6. CC Prediction and Relational Attributes Nr Target Recipient y xb xr’ xs xr xtr Nb Ns Nr-1 Ntr Thread Body Subject Other Relation Relation Words Words Recipients Thread Relations – e.g. Was a given recipient ever included on this email thread? Recipient Relationships – e.g. Does one of the other recipients report to the target recipient?

  7. CC-Prediction Learning in the Wild • As documents are added to Rexa, models of expertise for authors grows • As DEX obtains more contact information and keywords, organizational relations emerge • Model parameters can be adapted on-line • Priors on parameters can be used to transfer learned information between models • New relations can be added on-line • Modular model construction and intelligent model optimization enable these goals

  8. CC Prediction Upcoming work on Multi-Conditional Learning A discriminatively-trained topic model, discovering low-dimensional representations for transfer learning and improved regularization & generalization.

  9. Traditional, joint training (e.g. naive Bayes, most topic models) Traditional mixture model (e.g. LDA) Traditional, conditional training (e.g. MaxEnt classifiers, CRFs) Conditional mixtures (e.g. Jebara’s CEM, McCallum CRF string edit distance, ...) Multi-conditional(mostly conditional, generative regularization) Multi-conditional(for semi-sup) Multi-conditional(for transfer learning, 2 tasks, shared hiddens) Objective Functions for Parameter Estimation Traditional New, multi-conditional

  10. “Multi-Conditional Learning” (Regularization) [McCallum, Pal, Wang, 2006]

  11. Predictive Random Fieldsmixture of Gaussians on synthetic data [McCallum, Wang, Pal, 2005] Data, classify by color Generatively trained Multi-Conditional Conditionally-trained [Jebara 1998]

  12. Multi-Conditional Mixturesvs. Harmoniunon document retrieval task [McCallum, Wang, Pal, 2005] Multi-Conditional,multi-way conditionally trained Conditionally-trained,to predict class labels Harmonium, joint,with class labels and words Harmonium, joint with words, no labels

  13. DEX Beginning with a review of previous work, then new work on record extraction, with the ability to leverage new KBs in the wild, and for transfer

  14. Contact Info and Person Name Extraction Social Network Analysis Person Name Extraction Homepage Retrieval Keyword Extraction Name Coreference System Overview CRF WWW Email names

  15. An Example To: “Andrew McCallum” mccallum@cs.umass.edu Subject ... Search for new people

  16. Example keywords extracted Summary of Results Contact info and name extraction performance (25 fields) Expert Finding:When solving some task, find friends-of-friends with relevant expertise. Avoid “stove-piping” in large org’s by automatically suggesting collaborators. Given a task, automatically suggest the right team for the job. (Hiring aid!) Social Network Analysis:Understand the social structure of your organization.Suggest structural changes for improved efficiency.

  17. Importance of accurate DEX fields in IRIS • Information about • people • contact information • email • affiliation • job title • expertise • ... are key to answering many CALO questions... both directly, and as supporting inputs to higher-level questions.

  18. Extracted Record Name:Jane Smith, John Doe JobTitle: Professor, Administrative Assistant Company: U of California Department: Computer Science Phone: 209-555-5555, 209-444-4444 City: Boston Compatibility Graph .8 Jane Smith University of California 209-555-5555 .4 Computer Science -.5 Professor -.6 Boston Administrative Assistant -.5 -.4 University of California .4 John Doe 209-444-4444 Learning Field Compatibilities in DEX Professor Jane Smith University of California 209-555-5555 Professor Smith chairs the Computer Science Department. She hails from Boston, …her administrative assistant … John Doe Administrative Assistant University of California 209-444-4444

  19. Learning Field Compatibilities in DEX Extracted Record Professor Jane Smith University of California 209-555-5555 Professor Smith chairs the Computer Science Department. She hails from Boston, …her administrative assistant … John Doe Administrative Assistant University of California 209-444-4444 Name:Jane Smith, John Doe JobTitle: Professor, Administrative Assistant Company: U of California Department: Computer Science Phone: 209-555-5555, 209-444-4444 City: Boston Jane Smith University of California 209-555-5555 Professor Computer Science Boston Administrative Assistant University of California John Doe 209-444-4444

  20. Learning Field Compatibilities in DEX • ~35% error reduction over transitive closure • Qualitatively better than heuristic approach • Mine Knowledge Bases from other parts of IRIS for learning compatibility rules among fields • “Professor” job title co-occurs with “University” company • Area code / city compatibility • “Senator” job title co-occurs with “Washington, D.C” location • In the wild • As the user adds new fields & make corrections, DEX learns from this KB data • Transfer learning • between departments/industries

  21. Rexa A knowledge base of publications, grants, people, their expertise, topics, and inter-connections Learning for information extraction and coreference. Incrementally leveraging multiple sources of information for improved coreference Gathering information about people’s expertise and co-author, citation relations First a tour of Rexa, then slides about learning

  22. Previous Systems

  23. Previous Systems Cites Research Paper

  24. More Entities and Relations Expertise Cites Grant Research Paper Person Venue University Groups

  25. Learning in Rexa Extraction, coreferenceIn the wild: Re-adjusting KB after corrections from a user Also, learning research topics/expertise, and their interconnections

  26. Wide-spread interest, positive experimental results in many applications. Noun phrase, Named entity [HLT’03], [CoNLL’03]Protein structure prediction [ICML’04] IE from Bioinformatics text [Bioinformatics ‘04],… Asian word segmentation [COLING’04], [ACL’04]IE from Research papers [HTL’04] Object classification in images [CVPR ‘04] (Linear Chain) Conditional Random Fields [Lafferty, McCallum, Pereira 2001] Undirected graphical model, trained to maximize conditional probability of output sequence given input sequence where Finite state model Graphical model OTHERPERSONOTHERORGTITLE … output seq y y y y y t+2 t+3 t - 1 t t+1 FSM states . . . observations x x x x x t t +2 +3 t - t +1 t 1 input seq said Jones a Microsoft VP … (500 citations)

  27. IE from Research Papers [McCallum et al ‘99]

  28. IE from Research Papers Field-level F1 Hidden Markov Models (HMMs) 75.6 [Seymore, McCallum, Rosenfeld, 1999] Support Vector Machines (SVMs) 89.7 [Han, Giles, et al, 2003] Conditional Random Fields (CRFs) 93.9 [Peng, McCallum, 2004] error 40% (Word-level accuracy is >99%)

  29. Joint segmentation and co-reference Extraction from and matching of research paper citations. o s World Knowledge Laurel, B. Interface Agents: Metaphors with Character, in The Art of Human-Computer Interface Design, B. Laurel (ed), Addison-Wesley, 1990. c Co-reference decisions y y p Brenda Laurel. Interface Agents: Metaphors with Character, in Laurel, The Art of Human-Computer Interface Design, 355-366, 1990. Databasefield values c y c Citation attributes s s Segmentation o o 35% reduction in co-reference error by using segmentation uncertainty. 6-14% reduction in segmentation error by using co-reference. Inference: Variant of Iterated Conditional Modes [Wellner, McCallum, Peng, Hay, UAI 2004] see also [Marthi, Milch, Russell, 2003] [Besag, 1986]

  30. Rexa Learning in the Wildfrom User Feedback • Coreference will never be perfect. • Rexa allows users to enter corrections to coreference decisions • Rexa then uses this feedback to • re-consider other inter-related parts of the KB • automatically make further error correctionsby propagating constraints • (Our coreference system uses underlying ideas very much like Markov Logic, and scales to ~20 million mention objects.)

  31. Finding Topics in 1 million CS papers 200 topics & keywords automatically discovered.

  32. Topical Transfer Citation counts from one topic to another. Map “producers and consumers”

  33. Topical Diversity Find the topics that are cited by many other topics---measuring diversity of impact. Entropy of the topic distribution among papers that cite this paper (this topic). LowDiversity HighDiversity

  34. Some New Work onTopic Models Robustly capturing topic correlations Pachkinko Allocation Model Capturing phrases in topic-specific ways Topical N-Gram Model

  35. Pachinko Machine

  36. Pachinko Allocation Model [Li, McCallum, 2005] 11 Model structure, not the graphical model 21 22 Distributions over distributions over topics... Distributions over topics;mixtures, representing topic correlations 31 32 33 41 42 43 44 45 Distributions over words (like “LDA topics”) word1 word2 word3 word4 word5 word6 word7 word8 Some interior nodes could contain one multinomial, used for all documents. (i.e. a very peaked Dirichlet)

  37. “estimation” PAM 100 estimation bayesian parameters data methods estimate maximum probabilistic distributions noise variable variables noisy inference variance entropy models framework statistical estimating Topic Coherence Comparison “models, estimation, stopwords” “estimation, some junk” LDA 100 estimation likelihood maximum noisy estimates mixture scene surface normalization generated measurements surfaces estimating estimated iterative combined figure divisive sequence ideal LDA 20 models model parameters distribution bayesian probability estimation data gaussian methods likelihood em mixture show approach paper density framework approximation markov Example super-topic 33 input hidden units function number 27 estimation bayesian parameters data methods 24 distribution gaussian markov likelihood mixture 11 exact kalman full conditional deterministic 1 smoothing predictive regularizers intermediate slope

  38. Topic Correlations in PAM 5000 research paper abstracts, from across all CS Numbers on edges are supertopics’ Dirichlet parameters

More Related