Hi everyone ,
Iām trying to train a Hugging Face pretrained model for Telugu text summarization using PPO (Proximal Policy Optimization) with the trl
library, but Iām running into some issues.
with PPO config setup and library compatibility issues and training ,
another issue i was facing which is size mismatch .
please help me or give any example