This deep learning course provides a comprehensive introduction to attention mechanisms and transformer models the foundation of modern GenAI systems. Begin by exploring the shift from traditional neural networks to attention-based architectures. Understand how additive, multiplicative, and self-attention improve model accuracy in NLP and vision tasks. Dive into the mechanics of self-attention and how it powers models like GPT and BERT. Progress to mastering multi-head attention and transformer components, and explore their role in advanced text and image generation. Gain real-world insights through demos featuring GPT, DALL·E, LLaMa, and BERT.

Attention Mechanisms and Transformer Models Course

位教师:Priyanka Mehta
访问权限由 New York State Department of Labor 提供
您将学到什么
Apply self-attention and multi-head attention in deep learning models
Understand transformer architecture and its key components
Explore the role of attention in powering models like GPT and BERT
Analyze real-world GenAI applications in NLP and image generation
您将获得的技能
要了解的详细信息

添加到您的领英档案
7 项作业
了解顶级公司的员工如何掌握热门技能

积累特定领域的专业知识
- 向行业专家学习新概念
- 获得对主题或工具的基础理解
- 通过实践项目培养工作相关技能
- 获得可共享的职业证书

该课程共有2个模块
Explore the power of attention mechanisms in modern deep learning. Compare traditional neural architectures with attention-based models to see how additive, multiplicative, and self-attention boost accuracy in NLP and vision tasks. Grasp the core math and flow of self-attention, the engine behind Transformer giants like GPT and BERT and build a solid base for advanced AI development.
涵盖的内容
10个视频1篇阅读材料3个作业
Master multi-head attention and transformer models in this advanced module. Learn how multi-head attention improves context understanding and powers leading transformer architectures. Explore transformer components, text and image generation workflows, and real-world use cases with models like GPT, BERT, LLaMa, and DALL·E. Ideal for building GenAI-powered applications.
涵盖的内容
11个视频4个作业
获得职业证书
将此证书添加到您的 LinkedIn 个人资料、简历或履历中。在社交媒体和绩效考核中分享。
位教师

提供方
人们为什么选择 Coursera 来帮助自己实现职业发展

Felipe M.

Jennifer J.

Larry W.







