Learn to post-train and customize an LLM in this short course, “Post-training of LLMs ,” taught by Banghua Zhu, Assistant Professor at the University of Washington, and co-founder of NexusFlow.
Before a large language model can follow instructions or answer questions, it undergoes two key stages: pre-training and post-training. In pre-training, it learns to predict the next word or token from large amounts of unlabeled text. In post-training, it learns useful behaviors such as following instructions, tool use, and reasoning.
