Skip to content

yagmurilba/cs-552-modern-nlp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 

Repository files navigation

CS-552: Modern Natural Language Processing

Course Description

Natural language processing is ubiquitous in modern intelligent technologies, serving as a foundation for language translators, virtual assistants, search engines, and many more. In this course, we cover the foundations of modern methods for natural language processing, such as word embeddings, recurrent neural networks, transformers, and pretraining, and how they can be applied to important tasks in the field, such as machine translation and text classification. We also cover issues with these state-of-the-art approaches (such as robustness, interpretability, sensitivity), identify their failure modes in different NLP applications, and discuss analysis and mitigation techniques for these issues.

Quick access links:

Class

Platform Where & when
Lectures Wednesdays: 11:15-13:00 [STCC - Cloud C] & Thursdays: 13:15-14:00 [CE16]
Exercises Session Thursdays: 14:15-16:00 [CE11]
Project Assistance
(not every week)
Wednesdays: 13:15-14:00 [STCC - Cloud C]
QA Forum & Annoucements Ed Forum [link]
Grades Moodle [link]
Lecture Recordings Mediaspace [link]

All lectures will be given in person and live streamed on Zoom. The link to the Zoom is available on the Ed Forum (pinned post). Beware that, in the event of a technical failure during the lecture, continuing to accompany the lecture live via zoom might not be possible.

Recording of the lectures will be made available on Mediaspace. We will reuse some of last year's recordings and we may record a few new lectures in case of different lecture contents.

Lecture Schedule

Week Date Topic Suggested Reading Instructor
Week 1 19 Feb
20 Feb
Introduction | Building a simple neural classifier [slides]
Word embeddings
Suggested reading: Antoine Bosselut
Week 2 26 Feb
27 Feb
Classical LMs | Neural LMs: Fixed Context Models
Neural LMs: RNNs
Suggested reading: Antoine Bosselut
Week 3 5 Mar
6 Mar
Sequence-to-sequence Models | Transformers
Pretraining: GPT
Suggested reading: Antoine Bosselut
Week 4 12 Mar
13 Mar
[Online only] Pretraining: BERT | Transfer Learning
Pretraining: T5
Suggested reading: Antoine Bosselut
Week 5 19 Mar
20 Mar
Transfer Learning: Dataset Biases
Generation: Task
Suggested reading: - Antoine Bosselut
Week 6 26 Mar
27 Mar
Text Generation: Decoding & Training
Text Generation: Evaluation
Suggested reading: Antoine Bosselut
Week 7 / Week 8 2 Apr / 9 Apr
3 Apr / 10 Apr
Midterm / [Online only] LLMs: In-context Learning & Instruction Tuning
No Class / Project Description
Antoine Bosselut
Week 9 16 Apr
17 Apr
[Online only]Scaling laws | LLM Efficiency
No class (Work on your project)
Suggested reading: Antoine Bosselut
EASTER BREAK
Week 10 30 Apr
1 May
Ethics: Bias & Fairness | Toxicity & Disinformation
No class (Work on your project)
Suggested reading: Anna Sotnikova
Week 11 7 May
8 May
Tokenization | Multilingual LMs
No class (Work on your project)
Suggested reading: Antoine Bosselut
Week 12 14 May
15 May
Interpretability
No class (Work on your project)
Suggested reading: - Gail Weiss
Week 13 21 May
22 May
Retrieval-Augmented LLMs | Agents
No class (Work on your project)
Suggested reading: Antoine Bosselut
Week 14 28 May
29 May
Multimodality
Looking forward
Suggested reading: - Syrielle Montariol
Antoine Bosselut

Exercise Schedule

Week Date Topic Instructor
Week 1 20 Feb Intro + Setup [code] Beatriz Borges
Week 2 27 Feb LMs + Neural LMs: fixed-context models
Language and Sequence-to-sequence models
Simin Fan
Week 3 6 Mar Attention + Transformers, GPT Badr AlKhamissi
Week 4 13 Mar Pretraining and Transfer Learning Badr AlKhamissi
Week 5 20 Mar Transfer Learning Simin Fan
Week 6 27 Mar Generation Madhur Panwar
Week 7/8 MIDTERM / In-context Learning - GPT-3 Mete Ismayilzad

Exercises Session format:

  • TAs will provide a small discussion over the last week's exercises, answering any questions and explaining the solutions. (10-15mins)
  • TAs will present this week's exercise. (5mins)
  • Students will be solving this week's exercises and TAs will provide answers and clarification if needed.

Note: Please make sure you have already done the setup prerequisites to run the coding parts of the exercises. You can find the instructions here.

Grading:

Your grade in the course will be computed according to the following guidelines.

Submission Format

We will be using the Huggingface Hub as a centralized platform for submitting project artifacts, including model checkpoints and datasets. Please take some time to familiarize yourself with the functionalities of Huggingface Hub in Python to ensure a smooth workflow.

Late Days Policy

All milestones are due by 23:59 on their respective due dates. However, we understand that meeting deadlines can sometimes be challenging. To accommodate this, you will be given 2 late days for the semester to use at your discretion for group project milestones. Additionally, you will have 3 individual late days that can be applied to the individual project milestone. For group projects, if all members still have late days remaining, those days can be pooled and converted to group late days at a rate of one group late day per four individual late days. No extensions will be granted beyond the due dates, except for the final report, code, and data, which have a strict final deadline of June 8th. Late days will be automatically tracked based on your latest commit, so there is no need to notify us. Once all your late days are used, any further late submissions will incur a 25% grade deduction per day.

Midterm (30%):

More details will be announced in the next weeks.

Project (70%):

The project will be divided into 2 milestones and a final submission. Each milestone will be worth 15% of the final grade with the remaining 30% being allocated to the final report. Each team will be supervised by one of the course TAs or AEs.

More details on the content of the project and the deliverables of each milestone will be released at a later date.

Milestone 1:

  • Due: 4 May 2025

Milestone 2:

  • Due: 18 May 2025

Final Deliverable:

  • The final report, code, and date will be due on June 8th. Students are welcome to turn in their materials ahead of time, as soon as the semester ends.
  • Due: 8 June 2025

Contacts

Please email us at nlp-cs552-spring2025-ta-team [at] groupes [dot] epfl [dot] ch for any administrative questions, rather than emailing TAs individually. All course content questions need to be asked via Ed.

Lecturer: Antoine Bosselut

Teaching assistants: Angelika Romanou, Badr AlKhamissi, Beatriz Borges, Zeming (Eric) Chen, Simin Fan, Silin Gao, Mete Ismayilzada, Sepideh Mamooler, Madhur Panwar, Auguste Poiroux, Ayush Tarun

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published