Learn how to create, train, and tweak large language models (LLMs) by building one from the ground up! Master the complete process from initial design and creation, to pretraining on a general corpus, and fine-tuning for specific tasks.
Understanding transformer architecture, attention mechanisms, and basic language modeling
Building your first LLM from scratch, tokenization, and embedding strategies
Pretraining on large datasets, optimization techniques, and distributed training
Fine-tuning, instruction following, safety alignment, and deployment strategies
Proficiency in Python, object-oriented programming, and experience with data structures
Familiarity with PyTorch for deep learning, neural network training, and GPU programming
Understanding of basic ML concepts, neural networks, and deep learning fundamentals
Based on "Build a Large Language Model (From Scratch)" by Sebastian Raschka
Course launches soon!
Sign up for updates to be notified when registration opens.