This deep learning course provides a comprehensive introduction to attention mechanisms and transformer models the foundation of modern GenAI systems. Begin by exploring the shift from traditional neural networks to attention-based architectures. Understand how additive, multiplicative, and self-attention improve model accuracy in NLP and vision tasks. Dive into the mechanics of self-attention and how it powers models like GPT and BERT. Progress to mastering multi-head attention and transformer components, and explore their role in advanced text and image generation. Gain real-world insights through demos featuring GPT, DALL·E, LLaMa, and BERT.


您将学到什么
Apply self-attention and multi-head attention in deep learning models
Understand transformer architecture and its key components
Explore the role of attention in powering models like GPT and BERT
Analyze real-world GenAI applications in NLP and image generation
您将获得的技能
要了解的详细信息

添加到您的领英档案
June 2025
7 项作业
了解顶级公司的员工如何掌握热门技能

积累特定领域的专业知识
- 向行业专家学习新概念
- 获得对主题或工具的基础理解
- 通过实践项目培养工作相关技能
- 获得可共享的职业证书

该课程共有2个模块
Explore the power of attention mechanisms in modern deep learning. Compare traditional neural architectures with attention-based models to see how additive, multiplicative, and self-attention boost accuracy in NLP and vision tasks. Grasp the core math and flow of self-attention, the engine behind Transformer giants like GPT and BERT and build a solid base for advanced AI development.
涵盖的内容
10个视频1篇阅读材料3个作业
Master multi-head attention and transformer models in this advanced module. Learn how multi-head attention improves context understanding and powers leading transformer architectures. Explore transformer components, text and image generation workflows, and real-world use cases with models like GPT, BERT, LLaMa, and DALL·E. Ideal for building GenAI-powered applications.
涵盖的内容
11个视频4个作业
获得职业证书
将此证书添加到您的 LinkedIn 个人资料、简历或履历中。在社交媒体和绩效考核中分享。
位教师

提供方
从 Machine Learning 浏览更多内容
Google Cloud
Google Cloud
人们为什么选择 Coursera 来帮助自己实现职业发展




常见问题
The attention mechanism allows transformer models to focus on relevant parts of input sequences, weighing relationships between tokens to improve context understanding and accuracy in tasks like translation or text generation.
Yes, ChatGPT is built on the transformer architecture, specifically using a variant of the GPT (Generative Pre-trained Transformer) model, which enables it to generate human-like responses.
The Vision Transformer (ViT) applies self-attention to image patches instead of pixels, enabling the model to capture spatial relationships and global context for accurate image classification and understanding.
更多问题
提供助学金,