Wednesday, April 24, 2024

Lecture 14 (19/04/2024, 4h, TAs): homework 1b

 Introduction to homework 1b. In-class lab for the homework.

Lecture 13 (18/04/2024, 2h, S): Pre-trained language models

Pre-trained language models: BERT, GPT, RoBERTa, XLM.

Lecture 12 (12/04/2024, 3.5h): the Transformer

Introduction to the Transformer architecture. Encoder, decoder. Positional embeddings. Self-attention. Cross-attention. Decoding. Introduction to homework 1b.

Lecture 11 (11/04/2024, 2h): the attention mechanism

Neural language modeling. Context2vec. Neural language models with BiLSTMs. Contextualized word representations. Introduction to the attention. 

Tuesday, April 9, 2024

Lecture 10 (05/04/2024, 4h): notebooks on real-world reviewing example, training in NLP, hyperparameters, LSTMs

Notebooks on real-world review classification example, Part-of-Speech tagging brief introduction, LSTMs recap, Notebook on Part-of-Speech Tagging with LSTMs, data preprocessing and training procedure best practices.

 

Lecture 9 (04/04/2024, 2h): RNNs and Long-Short Term Memory Networks

Recurrent Neural Networks. Issues. Long-Short Term Memory Networks.

Friday, March 22, 2024

Lecture 8 (22/03/2024, 4h): introduction to language modeling

Negative sampling in the word2vec notebook. What is a language model? N-gram models (unigrams, bigrams, trigrams), together with their probability modeling and issues. Chain rule and n-gram estimation. Static vs. contextualized embeddings. Introduction to Recurrent Neural Networks.


Lecture 7 (21/03/2024, 2h): Word2vec notebook

PyTorch notebook on word2vec. More on homework 1.

Lecture 6 (15/03/2024, 4h): Negative sampling, homework 1 assignment

Negative sampling: the skipgram case; changes in the loss function. Homework 1 assignment.

Word2Vec Tutorial Part 2 - Negative Sampling · Chris McCormick

Thursday, March 14, 2024