Analyze & Deploy Scalable LLM Architectures is an intermediate course for ML engineers and AI practitioners tasked with moving large language model (LLM) prototypes into production. Many powerful models fail under real-world load due to architectural flaws. This course teaches you to prevent that.
即将结束: 只需 199 美元(原价 399 美元)即可通过 Coursera Plus 学习新技能。立即节省

您将获得的技能
- Kubernetes
- Containerization
- Model Deployment
- Configuration Management
- MLOps (Machine Learning Operations)
- Scalability
- Release Management
- Performance Testing
- Retrieval-Augmented Generation
- Infrastructure as Code (IaC)
- Application Deployment
- Continuous Delivery
- Performance Analysis
- Systems Analysis
- Cloud Deployment
- Analysis
- LLM Application
- Application Performance Management
- Large Language Modeling
- Performance Tuning
要了解的详细信息
了解顶级公司的员工如何掌握热门技能

积累特定领域的专业知识
- 向行业专家学习新概念
- 获得对主题或工具的基础理解
- 通过实践项目培养工作相关技能
- 获得可共享的职业证书

该课程共有3个模块
This module establishes the foundational mindset that "performance lives in the pipeline." Learners will discover that a large language model (LLM) application is a multi-stage system where overall speed is dictated by the slowest component. They will learn to deconstruct a complex Retrieval-Augmented Generation (RAG) architecture, trace a user request through it, and use system diagrams to form an evidence-based hypothesis about the primary performance bottleneck.
涵盖的内容
2个视频1篇阅读材料2个作业
In this module, learners move from hypothesis to evidence. They will learn to use system logging and profiling data to quantify the precise latency contribution of each stage in an LLM pipeline. The focus is on designing small, reversible, and hypothesis-driven experiments to prove or disprove their initial findings and distinguish a performance bottleneck's root cause from its symptoms.
涵盖的内容
1个视频2篇阅读材料2个作业
This module bridges the gap between a working prototype and a resilient, production-ready service. Learners will design and manage declarative deployments using Helm and Kubernetes, package a multi-component RAG stack, and implement Horizontal Pod Autoscaling (HPA) for dynamic, cost-efficient scaling. They will also master the critical operational skills of performing controlled, zero-downtime rollouts and rapid rollbacks.
涵盖的内容
2个视频2篇阅读材料2个作业
获得职业证书
将此证书添加到您的 LinkedIn 个人资料、简历或履历中。在社交媒体和绩效考核中分享。
位教师

提供方
从 Design and Product 浏览更多内容
状态:免费试用
状态:免费试用
状态:免费试用
人们为什么选择 Coursera 来帮助自己实现职业发展




常见问题
To access the course materials, assignments and to earn a Certificate, you will need to purchase the Certificate experience when you enroll in a course. You can try a Free Trial instead, or apply for Financial Aid. The course may offer 'Full Course, No Certificate' instead. This option lets you see all course materials, submit required assessments, and get a final grade. This also means that you will not be able to purchase a Certificate experience.
When you enroll in the course, you get access to all of the courses in the Specialization, and you earn a certificate when you complete the work. Your electronic Certificate will be added to your Accomplishments page - from there, you can print your Certificate or add it to your LinkedIn profile.
Yes. In select learning programs, you can apply for financial aid or a scholarship if you can’t afford the enrollment fee. If fin aid or scholarship is available for your learning program selection, you’ll find a link to apply on the description page.
更多问题
提供助学金,
¹ 本课程的部分作业采用 AI 评分。对于这些作业,将根据 Coursera 隐私声明使用您的数据。







