To do

  • Complete notes for lectures 12 (Kernels and SVM) and start writing notes for lectures 13
  • Start watching Andrew Ng lectures 17 (skip 16)
    • Learning With Large Datasets
    • Stochastic Gradient Descent
    • Mini-Batch Gradient Descent
    • Stochastic Gradient Descent Convergence
    • Online Learning
  • Read Karpathy blog post on reinforcement learning in preparation for Adam’s whiteboard lecture on it tomorrow

Done

  • Watched Andrew Ng lectures 17 up to and not including the end, which is about map reduce and data parallelism
  • Read Karpathy blog post on Policy Gradients and read the Nervana post on Q-Learning
    • All of this is under the field of Reinforcement Learning, which is what we’ll be using to compete in the upcoming VizDoom competition
  • Skipped writing notes for lecture 12 since Andrew Ng doesn’t do such a good job in covering them. Instead, I’m using this post on kernels as a learning resource

Extras