Date Topics Video from playlist
September 26th, 2025 Lecture 1: Transformer
[slides] [panopto]
• Background on NLP and tasks
• Tokenization
• Embeddings
• Word2vec, RNN, LSTM
• Attention mechanism
• Transformer architecture
Lecture 1
1:41:58
October 3rd, 2025 Lecture 2: Transformer-based models & tricks
[slides] [panopto]
• Attention approximation
• MHA, MQA, GQA
• Position embeddings (regular, learned)
• RoPE and applications
• Transformer-based architectures
• BERT and its derivatives
Lecture 2
1:47:19
October 10th, 2025 Lecture 3: Large Language Models
[slides] [panopto]
• Definition and architecture
• Mixture of experts
• Context length, temperature
• Sampling strategies
• Prompting, in-context learning
• Chain of thought
• Self-consistency
Lecture 3
1:48:44
October 17th, 2025 Lecture 4: LLM training
[slides] [panopto]
• Pretraining
• Quantization
• Hardware optimization
• Supervised finetuning (SFT)
• Parameter-efficient finetuning (LoRA)
Lecture 4
1:47:27
October 24th, 2025 Midterm
[exam] [solutions]
October 31st, 2025 Lecture 5: LLM tuning
[slides] [panopto]
• Preference tuning
• RLHF overview
• Reward modeling
• RL approaches (PPO and variants)
• DPO
Lecture 5
1:47:42
November 7th, 2025 Lecture 6: LLM reasoning
[slides] [panopto]
• Reasoning models
• RL for reasoning
• GRPO
• Scaling
Lecture 6
1:47:10
November 14th, 2025 Lecture 7: Agentic LLMs
[slides] [panopto]
• Retrieval-augmented generation
• Advanced RAG techniques
• Function calling
• Agents
• ReAct framework
Lecture 7
1:49:23
November 21st, 2025 Lecture 8: LLM evaluation
[slides] [panopto]
• LLM-as-a-judge overview
• Best practices and benefits
• Biases and pitfalls
Lecture 8
1:49:25
December 5th, 2025 Lecture 9: Current trends
[slides] [panopto]
• Recap
• Trending topics
• Closing thoughts
Lecture 9
1:51:31
December 10th, 2025 Final