470 likes | 666 Views
Shallow Processing: Summary. Shallow Processing Techniques for NLP Ling570 December 7, 2011. Roadmap. Looking back: Course units Tools and Techniques Looking forward Upcoming courses. Units #0,#1. Unit #0 (0.5 weeks): HW #1 Introduction to NLP & shallow processing Tokenization.
E N D
Shallow Processing:Summary Shallow Processing Techniques for NLP Ling570 December 7, 2011
Roadmap • Looking back: • Course units • Tools and Techniques • Looking forward • Upcoming courses
Units #0,#1 • Unit #0 (0.5 weeks): • HW #1 • Introduction to NLP & shallow processing • Tokenization
Units #0,#1 • Unit #0 (0.5 weeks): • HW #1 • Introduction to NLP & shallow processing • Tokenization • Unit #1 : • HW #2, #3 • Formal Languages and Automata (2.25 weeks) • Formal languages • Finite-state Automata • Finite-state Transducers • Morphological analysis
Unit #2 • Unit #2 (3.25 weeks): • HW #4, #5, #6, #7 • NgramLanguage Models and HMMs • NgramLanguage Models and Smoothing • Part-of-speech (POS) tagging: • Ngram • Hidden Markov Models
Units #3, #4 • Unit #3: Classification (3 weeks) • HW #8, #9 • Intro to classification & Mallets • POS tagging with classifiers • Chunking • Named Entity (NE) recognition
Units #3, #4 • Unit #3: Classification (3 weeks) • HW #8, #9 • Intro to classification & Mallets • POS tagging with classifiers • Chunking • Named Entity (NE) recognition • Unit #4: Selected Topics (1.5 weeks) • HW #10 • Clustering • Information Extraction • Summary
Main Techniques • Probability: • Chain rule:
Main Techniques • Probability: • Chain rule: • Bayes’ Rule:
Main Techniques • Formal languages: • Chomsky hierarchy
Main Techniques • Formal languages: • Chomsky hierarchy • Regular languages, regular expressions, regular grammars, finite state automata
Main Techniques • Formal languages: • Chomsky hierarchy • Regular languages, regular expressions, regular grammars, finite state automata • Regular relations, finite state transducers
Main Techniques • Formal languages: • Chomsky hierarchy • Regular languages, regular expressions, regular grammars, finite state automata • Regular relations, finite state transducers • Finite state morphology: two-level morphological analysis
Main Techniques • Formal languages: • Chomsky hierarchy • Regular languages, regular expressions, regular grammars, finite state automata • Regular relations, finite state transducers • Finite state morphology: two-level morphological analysis • Cascades of finite state transducers
Main techniques • Language Modeling and Hidden Markov Models
Main techniques • Language Modeling and Hidden Markov Models • N-gram language models • Maximum likelihood estimation
Main techniques • Language Modeling and Hidden Markov Models • N-gram language models • Maximum likelihood estimation • Smoothing • Laplace, Good-Turing, Backoff, Interpolation
Main techniques • Language Modeling and Hidden Markov Models • N-gram language models • Maximum likelihood estimation • Smoothing • Laplace, Good-Turing, Backoff, Interpolation • Hidden Markov Models • Markov assumptions • Forward algorithm & Viterbi decoding
Main Techniques • Classification for NLP:
Main Techniques • Classification for NLP: • Modeling NLP tasks as classification problems
Main Techniques • Classification for NLP: • Modeling NLP tasks as classification problems • Developing feature representations for instances
Main Techniques • Classification for NLP: • Modeling NLP tasks as classification problems • Developing feature representations for instances • Sequence labeling tasks and algorithms
Main Techniques • Classification for NLP: • Modeling NLP tasks as classification problems • Developing feature representations for instances • Sequence labeling tasks and algorithms • Beam search
Tools Developed • English tokenizer: HW#1 • FSA & FST acceptors: HW#2,#3 • FST morphological analyzer: HW#3 • N-gram language models with smoothing: HW#4,#5 • Authorship identification system: HW#5 • Hidden Markov Model: Training & Decoding: HW#6,7 • HMM POS Tagger • Classification-based text categorization, POS tagging • HW#8,#9 • Unsupervised POS tagger: HW#10
Corpora & Systems • Data:
Corpora & Systems • Data: • Penn Treebank • Wall Street Journal • Air Travel Information System (ATIS)
Corpora & Systems • Data: • Penn Treebank • Wall Street Journal • Air Travel Information System (ATIS) • Project Gutenberg • Federalist Papers, Jane Austen novels • Systems:
Corpora & Systems • Data: • Penn Treebank • Wall Street Journal • Air Travel Information System (ATIS) • Project Gutenberg • Federalist Papers, Jane Austen novels • Systems: • CARMEL Finite State Toolkit • Mallet Machine Learning Toolkit
Winter Courses • Ling571: Deep Processing Techniques for NLP • Parsing, Semantics (Lambda Calculus), Generation
Winter Courses • Ling571: Deep Processing Techniques for NLP • Parsing, Semantics (Lambda Calculus), Generation • Ling572: Advanced Statistical Methods in NLP • Roughly, machine learning for CompLing • Decision Trees, Naïve Bayes, MaxEnt, SVM, CRF,…
Winter Courses • Ling571: Deep Processing Techniques for NLP • Parsing, Semantics (Lambda Calculus), Generation • Ling572: Advanced Statistical Methods in NLP • Roughly, machine learning for CompLing • Decision Trees, Naïve Bayes, MaxEnt, SVM, CRF,… • Ling567: Knowledge Engineering for Deep NLP • HPSG and MRS for novel languages
Winter Courses • Ling571: Deep Processing Techniques for NLP • Parsing, Semantics (Lambda Calculus), Generation • Ling572: Advanced Statistical Methods in NLP • Roughly, machine learning for CompLing • Decision Trees, Naïve Bayes, MaxEnt, SVM, CRF,… • Ling567: Knowledge Engineering for Deep NLP • HPSG and MRS for novel languages • Ling575: Spoken Dialog Systems • Design, analysis, and implementation of SDS
Tentative Outline for Ling572 • Unit #0 (0.5 weeks): Basics • Introduction • Feature representations • Classification review
Tentative Outline for Ling572 • Unit #0 (0.5 weeks): Basics • Introduction • Feature representations • Classification review • Unit #1 (3 weeks): Classic Machine Learning • K Nearest Neighbors • Decision Trees • Naïve Bayes
Tentative Outline for Ling572 • Unit #3: (4 weeks): Discriminative Classifiers • Feature Selection • Maximum Entropy Models • Support Vectors Machines
Tentative Outline for Ling572 • Unit #3: (4 weeks): Discriminative Classifiers • Feature Selection • Maximum Entropy Models • Support Vectors Machines • Unit #4: (1.5 weeks): Sequence Learning • Conditional Random Fields • Transformation Based Learning
Tentative Outline for Ling572 • Unit #3: (4 weeks): Discriminative Classifiers • Feature Selection • Maximum Entropy Models • Support Vectors Machines • Unit #4: (1.5 weeks): Sequence Learning • Conditional Random Fields • Transformation Based Learning • Unit #5: (1 week): Other Topics • Semi-supervised learning,…
Ling572 Information • No required textbook: • Online readings and articles
Ling572 Information • No required textbook: • Online readings and articles • More math/stat content than 570 • Probability, Information Theory, Optimization
Ling572 Information • No required textbook: • Online readings and articles • More math/stat content than 570 • Probability, Information Theory, Optimization • Please try to register at least 2 weeks in advance
Beyond Ling572 • Machine learning: • Graphical models • Bayesian approaches • Online learning • Reinforcement learning • ….
Beyond Ling572 • Machine learning: • Graphical models • Bayesian approaches • Online learning • Reinforcement learning • …. • Applications: • Information Retrieval • Question Answering • Generation • Machine translation • ….
Ling 575: Spoken Dialog Systems • Design, analysis, and implementation of SDS • Will be offered online • Please make sure you have a microphone • Arrive early to test • Loew 202: T: 3:30-5:50
Notes • Grades should be submitted by 12/16 • Any issues (errors) with grades in Gradebook, please email by 12/14
Notes • Grades should be submitted by 12/16 • Any issues (errors) with grades in Gradebook, please email by 12/14 • Graduation Planning: • Students in Group 1 (planning to graduate ~8/12) • Due dates in early-mid January • 1st thesis proposal draft • 1st pre-internship report