Skip to the content.

KAIST AI605 - Deep Learning for NLP (Spring 2021)

Time & Location

Monday, Wednesday 2:30pm-3:50pm via Zoom (visit KLMS or email the instructor for the invitation)

Instructor

Minjoon Seo

minjoon@kaist.ac.kr

https://seominjoon.github.io

Office: KAIST Seoul Campus Building 9 Room 9202

Office Hours: Wed 4-5pm via email

TAs

Grading Policy

There is no exam in this class, and the grade will depend on the following criteria:

The final grade will be determined by the following policy:

Late Submission Policy

Course Description

This course covers recent advances in natural language processing area driven by deep learning. Topics include (but are not limited to)

There are four assignments: 3 coding and 1 writing.

There will be three coding assignments that involve training deep learning models for text classification (sentiment classification), token classification (machine reading comprehension), and text generation (machine translation). In the first two assignments, you will be asked to use PyTorch library only. In the third assignment, you will use a popular NLP tool, Hugging Face, to complete the assignment.

In this class, you will also learn how to write an NLP paper by analyzing the structure of recent papers published in NLP (ACL, EMNLP, NAACL) and machine learning (NeurIPS, ICLR, ICML) conferences. We will have in-class discussion where you will analyze frequent argument patterns in these papers, and the assignment will be writing a sample research paper that adopts an interesting pattern you found (with dummy experiments).

The final project is creating an open-domain question answering system on EfficientQA dataset. The final deliverable will be a report. However, if you are working on an NLP-related research project, you are welcome to work on it instead (but please consult with me first).

Textbook

There is no required textbook for this course but I highly recommend Speech & Language Processing, whose pdf version is available for free, for your reference.

Weekly Schedule (tentative)

  1. Introduction to NLP and and Review of Deep Learning
  2. Recurrent Neural Networks, Text Classification
  3. Token Classification
  4. Encoder-Decoder, Sequence Generation
  5. Transformer
  6. NLP paper writing
  7. Language Model
  8. Paper analysis presentation
  9. Pretrained Language Model & Finetuning
  10. NLP Tools
  11. Intro to the final project (Open-domain QA)
  12. Large Language Model
  13. Generalization & In-context Learning
  14. Final project presentation