In the lab for week2, one of the cells says “Training a fully fine-tuned version of the model would take a few hours on a GPU. To save time, download a checkpoint…” How was this checkpoint generated exactly? What were the training_args? What was the command to save the checkpoint? How long did it take on what kind of AWS instance?
Similarly, in the PEFT section of the lab, one of the cells says “To load a fully trained PEFT model, read a checkpoint of a PEFT model from S3.” Same questions here: How was this checkpoint generated exactly? What were the training_args? What was the command to save the checkpoint? How long did it take on what kind of AWS instance?
Thanks!