In this coding along video series, Sebastian Raschka, the author of "Build a Large Language Model from Scratch," guides viewers through teaching LLMs to follow instructions, focusing on building a small personal assistant similar to ChatGPT. He explains the process of fine-tuning a pre-trained model, emphasizing the importance of dataset preparation, including formatting data, tokenizing, and padding. The episode covers loading OpenAI weights, instruction fine-tuning, and evaluating the LLM's performance, highlighting the challenges in assessing free-form answers and introduces the use of OLAMA for automated evaluation. The discussion also extends to bonus materials like preference tuning with DPO and building a user interface for the fine-tuned model.
Sign in to continue reading, translating and more.
Continue