Link Search Menu Expand Document

Selected Topics in Deep Learning

Department of Mathematics, Sharif University of Technology. Fall 2023

This course offers an introductory exploration of deep learning, utilizing mathematical tools as its foundation. In the first half of the course, we revisit the core methods that underpin deep learning. Later on, we will cover a number of recent theoretical advances that aim to shed light on the mathematical foundations of deep learning.

Prerequisites

The following skills will be useful for success in this course:

  • Machine Learning: Some familiarity with machine learning will be helpful but not required; we will review important concepts that are needed for this course.

  • Programming: You should be comfortable programming in Python. You should be familiar with algorithms and data structures. Familiarity with numpy or similar frameworks for numeric programming will be helpful but is not strictly required.

  • Probability and Linear Algebra: You should have been exposed to probability distributions, random variables, expectations, etc.

Course Work

Grading will be based on:

  • Assignments (40%)
  • Final Exam (30%)
  • Paper Presentation (30%):

Project and Presentation

An essential component of this course involves delving deep into several papers related to specific research topics. This task is to be completed in pairs. While the choice of paper is based on your interests, it must relate to the theoretical foundations of deep learning. You are encouraged to consult with instructors and TAs when selecting your paper.

  • Proposal Submission: By the 1st of Azar, 1402, you are required to submit a 1-page proposal outlining your chosen research topic.

  • Presentation: During the final two weeks of the course, you will deliver a 20-minute presentation to succinctly summarize the research paper.

  • Paper Review: You are expected to submit a concise review of the paper. This review should provide an overview of the paper, its motivation, and the important theoretical results of the work. Additionally, critically evaluate its strengths and weaknesses. Furthermore, you are encouraged (although it is optional) to further explore the results or implement the ideas and compare them with other findings.

Tentative Lectures Schedule

Lecture Topic
Lecture 1 Introduction
Lectures 2-5 ML Overview
Lecture 6 Multi-layer Perceptrons
Lecture 7 Optimization
Lecture 8 Backpropagation
Lecture 9 Convolutional Neural Networks (CNN)
Lectures 10-11 Training DNN
Lecture 12 Recurrent Networks
Lecture 13-14 Attention, Transformers
Lecture 15 Large Language Models
Lecture 16 Score-based Generative Models
Lecture 17 Neural Tangent Kernels
Lecture 18 Bayesian Neural Networks
Lecture 19-27 Advanced Topics (TBA)
Lecture 28-29 Presentations