DeepLearning.AI
This course offers a deep dive into the main components of the transformer architecture that powers large language models (LLMs). Gain a strong technical foundation in transformers, understand recent improvements, and explore implementations in the Hugging Face Transformers library.
In this course, you'll learn how a transformer network architecture that powers LLMs works. You'll build the intuition of how LLMs process text and work with code examples that illustrate the key components of the transformer architecture. By the end of this course, you’ll have a deep understanding of how LLMs process language and you’ll be able to read through papers describing models and understand the details that are used to describe these architectures.
AI Enthusiasts
Individuals interested in understanding the inner workings of transformer architectures that power today's LLMs.
Data Scientists
Professionals looking to deepen their knowledge of transformer models and their applications in AI.
Developers
Developers aiming to build applications using large language models and understand their underlying architecture.
Gain a deep understanding of transformer architectures that power today's LLMs. Learn key components like tokenization, embeddings, and self-attention, and explore recent improvements. Ideal for AI enthusiasts, data scientists, and developers looking to advance their careers.
1 / 3
Basic understanding of machine learning concepts
Familiarity with programming languages such as Python
Interest in AI and language models
Jay Alammar
Director and Engineering Fellow, Cohere
Co-author of Hands-On Large Language Models
Maarten Grootendorst
Senior Clinical Data Scientist, Netherlands Comprehensive Cancer Organization
Cost
Free
Duration
Dates
Location