

Inference techniques for local and cloud LLM deployment
包含在 中
推荐体验
您将学到什么
The principles of LLM inference and prompt pipelines for real-world tasks.
Running small and medium LLMs locally with Ollama and deploying larger models in the cloud using Python.
Building and documenting LLM-powered tools ready for real-world use.
要了解的详细信息

添加到您的领英档案
了解顶级公司的员工如何掌握热门技能

积累 Software Development 领域的专业知识
- 向行业专家学习新概念
- 获得对主题或工具的基础理解
- 通过实践项目培养工作相关技能
- 通过 Meta 获得可共享的职业证书

从 Software Development 浏览更多内容
人们为什么选择 Coursera 来帮助自己实现职业发展




常见问题
Developers, entrepreneurs, and technical professionals with 1-2 years of Python knowledge who want to build AI-enabled assistants. Ideal for those looking to upskill in generative AI development or create practical business solutions using Llama models.
1-2 years of Python programming experience (for those who need to meet this prerequisite, start with the Meta Programming in Python course)
Familiarity with command-line interfaces
Understanding of basic software development concepts
Basic knowledge of REST APIs
In this program, you will be guided to access the Llama 4 Scout 17B and Llama 3.1 8B models via API in Courses 1 and 2. The course content includes examples using one of the API providers, but you are free to choose any provider that offers access to Llama models for your learning experience. Examples of such providers include Together AI, Groq, Hugging Face, and others.
In Course 3, you will be guided to use the Llama 3.1 8B model in a local environment. Llama models are available from multiple sources, with the best place to start being https://www.llama.com.
Models are also hosted and distributed by partners such as Amazon Web Services, Microsoft Azure, Google Cloud, IBM Watsonx, Oracle Cloud, Snowflake, Databricks, Dell, Hugging Face, Groq, Cerebras, SambaNova, and many others. See the Llama.com FAQ for more information.
更多问题
提供助学金,