Computational Linguistics

CMSC 723 / LING 723 / INST 735

Jan 26: Welcome to CL1!

Required Lectures:

Optional Readings:

Optional Lectures (Stuff you should already know):

Jan 28: Review

Required Lectures [slides]:

Required Readings (skim if you know this already):

Also of interest:

Feb 2: Historical Background and the NLP Pipeline

Required Lectures:

A famous quote: "Je n’ai fait celle-ci plus longue que parce que je n’ai pas eu le loisir de la faire plus courte." And another: "Il meglio è l'inimico del bene". Such is the case with this week's video lectures. Hopefully the slightly greater length will be worthwhile in terms of providing you with interesting things to think about and discuss in class.

Some things to think about for class:

  • What levels of analysis are being illustrated in the Alexa examples?

  • In terms of those levels, where is Alexa doing well, or poorly, or not even trying?

  • Based on what you've seen so far, how does dealing with language computationally compare/contrast with other computational problems you've looked at -- e.g. vision, planning, gene sequencing, cryptography, ...?

Required Readings:

Other readings possibly of interest:

  • Tenney, Das, and Pavlick (2019), BERT Rediscovers the Classical NLP Pipeline. This reading is not intended for general class discussion but may be of interest to students who have already encountered deep learning models applied to text.

Feb 5 Homework: Limericks


Feb 9: Representation Learning

Required Lectures:

Required Readings:

Feb 11: Deep Learning

Required Lectures:

Required Readings:

Feb 12 Homework: Logistic Regression


Feb 16: Words, Words, Words

Required Lectures:

Required Readings:

Also potentially of interest:

iconv -f windows-1252 -t utf-8 -c

-c When this option is given, characters that cannot be converted are silently discarded, instead of leading to a conversion error.

-f encoding, --from-code=encoding

Specifies the encoding of the input.

-t encoding, --to-code=encoding

Specifies the encoding of the output.



Feb 18: Pytorch

Required Lectures:

Required Readings:

Feb 23: Word Meaning

Required Lectures:

Required Readings:

Not required

  • (We were originally going to include SLP 20 for this lecture but will not)

To think about

  • Without looking in a dictionary, what meanings would you enumerate for the verb break ?

Feb 25: Sequential Structure

Required Lectures:

Note that slides I showed in these lectures are linked with Chapter 3 in the SLP book draft.

Required Readings:

  • Pinker, pp. 89-97

  • SLP 3

  • SLP 8 through 8.4

Mar 2: Sequential Structure, continued

Required Lectures:

Required Readings:

  • SLP Appendix A (with a focus on the forward-backward algorithm)

  • Pinker 97-103

Mar 3 Homework: Pytorch Logistic Regression


Mar 4: Syntactic Structure

Required Lectures:

If these links to Panopto do not work (they worked for me, but someone has reported an error), the videos (.mp4) should now be downloadable by anyone from a UMD account at https://umd.box.com/s/ypl9lrv6u462ldxs9h6ocmzf5lgo38gq.

Required Readings:

  • SLP 12

  • Pinker 103-125

Mar 9: Syntactic Structure, continued

Required Lectures:

Required Readings:

  • SLP 13

    • Main focus is the start of the chapter through Section 13.2

    • In Section 13.3, focus on the discussion of Equations 13.6-13.10 (and how they relate to CKY)

    • Section 13.4 will be covered in a later lecture on evaluation; you can just skim if you like

    • Read Section 13.5

    • Read Sections 13.6.3-13.6.4, with a focus on the idea of agenda-driven parsing (and again the relationship to CKY)

Recommended Readings:

  • SLP 14

    • Our main focus is on the idea of dependency representations, not dependency parsing algorithms

Recommended ahead of class

Exercises for class


Mar 11: Sentence Meaning

Required Lectures:

Required Readings:

Mar 13 Homework: Project Proposal (EARLY)


Mar 16-18: Spring Break

Have fun!

Mar 23: Evaluation of NLP Systems

Required Lectures:


Required Readings:




Mar 24 Homework: Project Proposal (LAST CHANCE)


Mar 25: Pre-midterm discussion

Take-home midterm to be distributed at end of class


Apr 8: Non-neural ML for sequences - Structured Perceptron & PYLM

Required Lectures:

Required Readings:


APR 9 Homework: Bigrams


Apr 13: Named Entities / Entity Linking / Coreference

Required Lectures:

Required Readings:

Apr 15: Machine Translation

Required Lectures:

Required Readings:

Also of interest:



APR 16 Homework: Project DELIVERABLE


Apr 20: Ethical AI

Required Lectures:

Required Readings:

Recommended



Apr 27: NLP and Computational Social Science

Required Lectures:

Required Readings:

  • SLP Ch 20 sections up to Section 20.3; Section 20.5.1


APR 30 Homework: Named Entity Recognition

Last homework!

May 6: Linguistics as a Scientific Endeavor

This will be a live lecture for half the class time, followed by discussion. Readings below are optional and not required.

Recommended to view ahead of time (<20min):

Recommended Readings


May 11: Project Presentations

Post your presentation to Piazza (we'll provide a thread) by midnight May 9, and we'll use the time in class to ask questions.