Overview

Recent developments in neural networks (aka “deep learning”) have drastically advanced the performance of machine perception systems in a variety of areas including computer vision, robotics, and human shape modeling This course is a deep dive into deep learning algorithms and architectures with applications to a variety of perceptual and generative tasks.


Announcements

04.05.2023
The Machine Perception lectures on 10. / 11. May will be fully remote. Please join and ask your questions on Zoom. Exercise classes take place as usual.
03.05.2023
The mock exam is now available here.
14.04.2023
The Machine Perception lectures on 19. / 20. April will be fully remote. Please join and ask your questions on Zoom. Exercise classes take place as usual.
28.03.2023
The lectures this week are cancelled. Please use last years recordings (part I and part II) to study the material. The tutorial will still take place on Thursday. There won't be tutorials this week.
23.02.2023
Project descriptions have been added here!
11.01.2023
More info coming soon!

Learning Objectives

Students will learn about fundamental aspects of modern deep learning approaches for perception and generation. Students will learn to implement, train and debug their own neural networks and gain a detailed understanding of cutting-edge research in learning-based computer vision, robotics, and shape modeling. The optional final project assignment will involve training a complex neural network architecture and applying it to a real-world dataset.

The core competency acquired through this course is a solid foundation in deep-learning algorithms to process and interpret human-centric signals. In particular, students should be able to develop systems that deal with the problem of recognizing people in images, detecting and describing body parts, inferring their spatial configuration, performing action/gesture recognition from still images or image sequences, also considering multi-modal data, among others.

We will focus on teaching: how to set up the problem of machine perception, the learning algorithms, network architectures, and advanced deep learning concepts in particular probabilistic deep learning models.
The course covers the following main areas:
I) Foundations of deep learning.
II) Advanced topics like probabilistic generative modeling of data (latent variable models, generative adversarial networks, auto-regressive models, invertible neural networks).
III) Deep learning in computer vision, human-computer interaction, and robotics.


Schedule

Subject to change. Materials only available from within ETH network.

Wk.Date ContentMaterial Exercise Session
1 22.02
Deep Learning Introduction

Class content & admin

slides
1 23.02
-- No Class --
2 01.03
02.03
Training Neural Networks

Backpropagation
Feedforward Networks,
Representation Learning

slides pt. I

slides pt. II
Perceptron Visualization Notebook

Tutorial Implement your own MLP

slides
XOR Notebook
XOR Solutions
Eye-Gaze Notebook
Eye-Gaze Solutions

Tutorial Linear Regr.

slides
Linear Regression Notebook

Pen & Paper Backprop.

exercise
exercise solution
3 09.03.
10.03.
Convolutional Neural Networks

slides pt. I
slides pt. II

Additional material:
RSVP
Cortical Neuron -->

Tutorial CNNs in Pytorch

slides
CNN Notebook

Pen & Paper CNN

exercise
exercise solution
4 15.03.
Fully Convolutional Neural Networks
slides
4 16.03.
Recurrent Neural Networks

LSTM, GRU, Backpropagation through time

slides

Tutorial RNNs in Pytorch

slides
RNN Notebook

Pen & Paper RNN

exercise
exercise solution
5 23.03.
24.03
Generative Models Pt. I: Latent Variable Models

Variational Autoencoders, etc.

slides pt. I
slides pt. II

Class Tips for Training I

slides

Pen & Paper VAE

exercise
exercise solution
6 30.03.
31.03.
Generative Models Pt. II: Autoregressive Models

PixelCNN, PixelRNN, WaveNet, Stochastic RNNs

slides pt. I
slides pt. II

Class Tips for Training II

slides

Pen & Paper AR

exercise
exercise solution
7 05.04.
06.04.
Generative Models Pt. III: Normalizing Flows and Invertible Neural Networks

slides NF Pt. I
slides NF Pt. II

Pen & Paper NF

exercise
exercise solution
8 12.04.
23.04.
-- No Class (Easter) --
9 19.04.
20.04.
Generative Models Pt. IV: Implicit Models

Generative Adversarial Networks & Co

slides GAN Pt. I
slides GAN Pt. II

Tutorial Exercise Discussion and Euler

slides PP Backprop&CNN
slides Euler

Pen & Paper GAN

exercise
exercise solution
10 26.04
27.04
Generative Models Pt. V: GAN Applications and Diffusion Models
slides GAN Pt. III
slides Diffusion Models

Pen & Paper class Diffusion Models

exercise
exercise solution
11 03.05
04.05
Implicit Surfaces and Neural Radiance Fields
slides NIR Pt. I
slides NIR Pt. II

Tutorial Exercise Discussion AR

slides

Pen & Paper class Implicit Surfaces

exercise
exercise solution
12 10.05.
11.05.
Parametric Human Body Models and Applications

slides PBM

Tutorial Exercise Discussion GAN & NF

slides

Pen & Paper PBM

exercise
exercise solutions
13 17.05.
18.05.
-- No classes or exercise sessions --
13 24.05.
25.05.
Reinforcement Learning
slides RL pt. I
slides RL pt. II

Pen & Paper RL

exercise
exercise solutions
14 01.06
AIT Open House
07.06
Exam (13:30 - 16:30; HIL E 3, HIL E 4)
27.07 / 28.07
Exam Review (14:00 - 15:00; Location ML F 39)

Exercise Sessions

Please refer to the above schedule once available for an overview of the planned exercise slots. We will have three different types of activities in the exercise sessions:

  1. Tutorial: Interactive programming tutorial in Python taught by a TA. Code will be made available.
  2. Class: Lecture-style class taught by a TA to give you some tips on how to train your neural network in practice.
  3. Pen & Paper: Pen & paper exercises that are not graded but are helpful to prepare for the written exam. Solutions will be published on the website a week after the release and discussed in the exercise session if desired.


Project

Overview

There will be a multi-week project that gives you the opportunity to have some hands-on experience with training a neural network for a concrete application.

The project grade will be determined by two factors: 1) a competitive part based on how well your model fairs compared to your fellow students' models and 2) the idea/novelty/innovativeness of your approach based on a written report to be handed in by the project deadline. For each project there will be baselines available that guarantee a certain grade for the competitive part if you surpass them. The competition will be hosted on a online platform - more details will be announced here.

Check out the project descriptions here (you will need to log in with your ETH LDAP).


Exam

To give you a rough idea what to expect for the exam, we release a mock exam which you can download here:


Registration as Non-primary Target Group

Registrations have been closed.