Today, I'm launching my first course, all about Autograd.
Here is a brief content description:
The course comprises five video lectures (~4.5 hours total) with Jupyter notebooks and exercises. I have set up a private zulip instance (see below), so you can also ask on the forums, and I will do regular Q&A sessions (approximately every 3 weeks depending on demand).
You are an intermediate to advanced PyTorch user and want to know everything about autograd in Python. This course tries to cover autograd in full detail, so
- if you ever spent a day or two debugging why it said
Trying to backward through the graph a second time, but the buffers have already been freed.,
- built a custom
autograd.Functionand wondered why the
gradcheckworked, but the
gradgradcheckdid not, or
- want to know about these for making your work more efficient,
this course is for you.
You will learn:
- How autograd keeps track of your computation for autodifferentiation,
- common error messages, their causes, and how to debug them,
- all about defining your own
- how higher order derivatives work in PyTorch,
- other APIs (functional, forward-mode autograd),
- autograd graph debugging with multiple levels of detail,
- finding out how exactly inference mode works by using a debugger and the PyTorch source code.
Price: EUR 60.
What former participants say:
I was working on PyTorch for the last two years, but this course just made things so much interesting and clear for me. Kudos to Thomas for providing me this opportunity to learn.
Edit (Sep 27): Registration for the second batch is open until Sep 30. Note that you need to navigate the chat to #general / courses to purchase the course.