Hello,
I am trying to repeat the lab 2 fine-tuning exercise using peft approach on my own Mac mini. I am using flan-T5-small so that the model can fit in the 16G memory. I increased max_step to 10000 and I increased no of epochs to 5. The training finished after 24 hours. However, I only observed 0% - 3% improvement of the rouge scores by the peft model against the original model. Could anybody tell me what I can do to obtain the similar level of improvement as demonstrated in the Lab 2 vedio ?
Here are my peft configurations and training configurations:
lora_config = LoraConfig(r=32, lora_alpha=32, target_modules=[‘q’, ‘v’, ‘k’, ‘o’],
lora_dropout=0.05, bias=‘lora_only’,task_type=TaskType.SEQ_2_SEQ_LM)
peft_training_args=TrainingArguments(
output_dir=output_dir,
learning_rate=1e-4, # Better LR for small model
num_train_epochs=5, # Proper epoch-based training
max_steps=10000,
logging_steps=100,
report_to=“tensorboard”,
)
Thanks a lot!
Quan
When creating a post, please add:
- Week # must be added in the tags option of the post.
- Link to the classroom item you are referring to:
- Description (include relevant info but please do not post solution code or your entire notebook):