This schedule is tentative and subject to change as the semester progresses.
Date | Lecture | Reading | Assignment | ||
---|---|---|---|---|---|
Introduction, N-grams | |||||
Th | 1/12 | 1 | What is NLP? | SLP2 ch. 1 | A0: The Basics & A0.5: ChatGPT |
Tu | 1/17 | 2 | Why do we need data?; Basic Text Processing (Regular Expressions, Normalization) | Working with text in Python 3, SLP3 2.0–2.3, NLTK Book 3.2–3.10 | |
Th | 1/19 | 3 | N-gram language models | SLP3 3.0–3.5.2; Goldwater probability review through section 4 | A0 due (A0.5 due the following day) |
Tu | 1/24 | 4 | N-gram language models contd. | A1: Language Models | |
Th | 1/26 | 5 | Overview of Linguistics | Features English is missing - but most other languages have; SLP2 1.1, SLP3 13.1 | |
Classification | |||||
Tu | 1/31 | 6 | Classification: naïve Bayes; Noisy Channel Model | SLP3 4.0–4.5 | |
Th | 2/2 | 7 | Lexical semantics: senses, relations, classes | SLP3 23.0-23.3 | |
Fr | 2/3 | A1 due | |||
Tu | 2/7 | 8 | Linear models for classification: features & weights | SLP3 5.0–5.2 | A2: Perceptron |
Th | 2/9 | 9 | Linear models for classification: discriminative learning (perceptron, SVMs, MaxEnt) | Daumé The Perceptron: 4.0–4.3; SLP3 4.7–4.8, ch. 5 (Further readings are suggested in slides) | |
Tu | 2/14 | 10 | Linear models contd. | ||
Sequential Prediction | |||||
Th | 2/16 | 11 | Parts of speech; Review | SLP3 8.0–8.3 | |
F | 2/17 | A2 due | |||
Tu | 2/21 | No class: classes follow Monday schedule | |||
Th | 2/23 | 12 | POS tagging: HMMs | SLP3 8.4.0–8.4.3 Eisenstein Notes, 7.1: Part-of-speech tagging | |
F | 2/24 | GRADED QUIZ 1 | (See study guide worksheet) | ||
Tu | 2/28 | 13 | Algorithms for HMMs (mainly Viterbi); BONUS: Discriminative tagging with the structured perceptron | SLP3 8.4.4–8.4.6, Appendix A; BONUS: Eisenstein Notes, 7.5 (no need to read beyond structured perceptron); Neubig slides | |
Th | 3/2 | 14 | Annotation; Universal POS annotation activity (tagset) | A3: HMM | |
M-F | 3/6-3/10 | No class or office hours: Spring Break | |||
Distributed Representations and Neural Networks | |||||
Tu | 3/14 | 15 | Distributional representations and similarity | SLP3 ch. 6 | |
Th | 3/16 | 16 | Deep learning and neural networks | SLP3 7.0–7.1, 7.3–7.5 | |
F | 3/17 | A3 due | |||
Tu | 3/21 | 17 | Neural sequence modeling with RNNs | SLP3 9.1–9.8 | A4: LSTMs |
Hierarchical Sentence Structure | |||||
Th | 3/23 | 18 | English syntax, CFGs | SLP3 17.0–17.2, skim 17.3 | |
Tu | 3/28 | 19 | Syntax contd.; project: review task options and submit preferences | ||
Th | 3/30 | 20 | (P)CFG parsing: Parsing as search; CNF. | P0: Submit project team with topic | |
F | 3/31 | ||||
M | 4/3 | A4 due | |||
Tu | 4/4 | 21 | (P)CFG parsing contd.: CKY walkthrough | SLP3 17.5–17.8, Appendix C C.0-C.4 | P1: 1-2 page proposal due |
Th-M | 4/6-4/10 | No class or office hours: Easter Break | |||
Tu | 4/11 | 22 | Neural sequence-to-sequence models | SLP3 9.7–9.8, 10.0–10.2 | |
Th | 4/13 | 23 | In-class project work session and team meetings with course staff | ||
M | 4/17 | A5: Syntax | |||
Tu | 4/18 | 24 | Dependency parsing (Tatsuya Aoyama) | SLP3 18.0-18.2, 18.4 | P2: Progress update, including lit review, due |
Sequence-to-Sequence, Translation, and Other Applications | |||||
Th | 4/20 | 25 | Text generation (mainly QA, summarization, MT; mention dialog, image captioning) (Shabnam Behzad) | SLP3 13, 14.2, 14.4–14.7 | |
Tu | 4/25 | 26 | Guest Lecture by Jonathan Kummerfeld: semantic parsing, explainability, interaction | TBA | |
Th | 4/27 | 27 | Context in language processing + Wrap-up | ||
F | 4/28 | A5 due | |||
Tu | 5/2 | 28 | Project presentations | ||
F | 5/5 | GRADED QUIZ 2 | Study guide to appear | ||
F | 5/12 | No meeting during finals | PROJECT REPORT DUE | ||
Visit the GUCL website for NLP talks this semester and beyond! |