True, He also uses the term g(theta), which seemingly he meant for a function that computes gradient , rather than the usual notation of activation function.
I think either they re-used a video from some other playlist to use here or the course is actually missing a video.