Let’s get started by importing the modules we need for this notebook as well as setting a few hyperparameters that are needed throughout the notebook.

Tips:

This notebook will use the same hyperparameters as were used in Part 1:

  • Learning Rate (lr): how much model parameters are updated at each batch/epoch
  • Batch Size: number of data points used to estimate gradients at each iteration
  • Epochs: Number of times to iterate over our entire dataset in optimization process

and are defined below:

 
© Chishiki-AI  |   Cornell University    |   Center for Advanced Computing    |   Copyright Statement    |   Access Statement
CVW material development is supported by NSF OAC awards 1854828, 2321040, 2323116 (UT Austin) and 2005506 (Indiana University)