This course addresses the challenge of machine learning (ML) in the context of small datasets, a significant issue due to ML's increasing data demands. Despite ML's success in various fields, many areas can't provide large labeled datasets because of costs, privacy, or security laws. As big data becomes standard, efficiently learning from smaller datasets is crucial. This course, ideal for graduate students with some ML experience, focuses on modern deep learning techniques for small data applications relevant in healthcare, military, and various industry sectors. Prerequisites include ML familiarity and Python proficiency. Deep learning experience is not necessary but beneficial.
通过 Coursera Plus 提高技能,仅需 239 美元/年(原价 399 美元)。立即节省

您将获得的技能
要了解的详细信息

添加到您的领英档案
8 项作业
了解顶级公司的员工如何掌握热门技能

该课程共有7个模块
In this module, we will explore the pivotal role of data as the foundation for machine learning algorithms. We begin by discussing the significance of large datasets in training deep learning models as these datasets are crucial for the models’ successful application and effectiveness. We will also delve into the challenges associated with small datasets, particularly in sensitive fields such as healthcare and defense, where data acquisition is often difficult, costly, or subject to stringent privacy and security regulations. To address these challenges, the course will introduce various strategies for making the most of limited data, including data-efficient machine learning techniques and the use of synthetic data augmentation. Additionally, we will present the course structure and discuss a curated selection of research papers that align with and enrich our course topics.
涵盖的内容
2个视频13篇阅读材料1个作业
2个视频• 总计16分钟
- Data Matters• 8分钟
- Setting Up Your Local Environment• 8分钟
13篇阅读材料• 总计81分钟
- Course Overview• 1分钟
- Syllabus - Machine Learning for Small Data• 10分钟
- Academic Integrity• 1分钟
- Data Matters—Especially for Deep Learning• 2分钟
- Data-Parameters-Power Scaling in AI Model• 5分钟
- Exponential Growth of Training Data• 10分钟
- Exponential Growth of Model Complexity• 5分钟
- Exponential Growth in Computational Resources• 5分钟
- The Scale Paradox: When Smaller ML Models Outperform Giants• 5分钟
- Large Datasets for Deep Learning• 10分钟
- What is Small Data?• 2分钟
- Installing PyTorch• 5分钟
- Large vs. Small Datasets in Machine Learning• 20分钟
1个作业• 总计10分钟
- Module 1 Quiz• 10分钟
In this module, we will delve into the core aspects of machine learning with a focus on the importance of data, particularly in deep learning applications. We start by emphasizing how large datasets are essential for training deep learning models effectively, as they enable the models to capture and learn from complex patterns, improving their overall performance. Additionally, we'll explore the intersection of data availability, computational power, and model capacity, highlighting how these elements interact to refine model accuracy and efficiency. Furthermore, the module will cover computing advancements beyond Moore's Law and their impact on machine learning, illustrating how modern hardware like CPUs, GPUs, and TPUs enhance computational capabilities critical for training sophisticated models. We'll also delve into scaling laws in deep learning, discussing empirical findings that show how model performance improves predictably with increases in dataset size and model complexity, although with diminishing returns. To provide a deeper theoretical foundation, we'll examine the Vapnik-Chervonenkis (VC) theory, which offers insights into how learning curves and model complexity relate to a model’s ability to generalize from training data. This discussion will extend to practical applications and theoretical limitations, helping to frame machine learning challenges in terms of data sufficiency, model fitting, and the balance between bias and variance. By the end of this module, students will have a thorough understanding of the dynamic interplay between these factors and their implications for machine learning practice and research.
涵盖的内容
1个视频19篇阅读材料2个作业1个应用程序项目
1个视频• 总计9分钟
- Machine Learning Model Performance• 9分钟
19篇阅读材料• 总计144分钟
- Ingredients Relationship• 10分钟
- Computing Power: Growth Beyond Moore’s Law• 10分钟
- Scaling Laws• 5分钟
- Learning Curves• 15分钟
- Model Capacity Required to Fit Data• 3分钟
- Model Performance and Dataset Size• 2分钟
- Model Performance and Model Capacity• 2分钟
- Bias-Variance Trade-Off• 15分钟
- From a Linear Algebra Perspective• 2分钟
- Underdetermined Problems and Overparameterized Models• 8分钟
- Revisiting Bias-Variance with Double Descent• 8分钟
- Comparison of Learning Paradigms• 15分钟
- A Learning Machine• 2分钟
- How Do We Characterize Model Complexity?• 1分钟
- Vapnik–Chervonenkis (VC) Dimension - Shattering• 10分钟
- Notions of VC Dimension• 10分钟
- Examples of Shattering and VC Dimension• 10分钟
- VC Dimension in Neural Networks• 15分钟
- Resources• 1分钟
2个作业• 总计60分钟
- Calculating the VC Dimension of SVM Models• 30分钟
- Module 2 Quiz• 30分钟
1个应用程序项目• 总计10分钟
- Examples of Learning Machines• 10分钟
In this module, we’ll explore transfer learning and its role in data-efficient machine learning, where models leverage knowledge from previous tasks to improve performance on new, related tasks. We’ll also cover various types of transfer learning, including transductive, inductive, and unsupervised methods, each addressing different challenges and applications. We’ll discuss some practical steps for implementing transfer learning, such as selecting and fine-tuning pre-trained models, to reduce reliance on large datasets. We’ll also examine data-driven and physics-based simulations for data augmentation, highlighting their use in enhancing training under constrained conditions. Finally, we’ll review key papers on transfer learning techniques to address data scarcity and improve model performance.
涵盖的内容
1个视频15篇阅读材料1个作业
1个视频• 总计6分钟
- Transfer Learning• 6分钟
15篇阅读材料• 总计72分钟
- Data-efficient Machine Learning• 10分钟
- Leveraging Pre-trained Models for Efficient Machine Learning• 2分钟
- Vanilla Transfer Learning • 2分钟
- Types of Transfer Learning• 2分钟
- Transductive Transfer Learning Algorithms• 10分钟
- Inductive Transfer Learning Algorithms• 10分钟
- Transductive Examples I• 5分钟
- Transductive Examples II• 5分钟
- Transductive Examples III• 5分钟
- Inductive Examples• 5分钟
- Multi-Task Learning & Meta-Learning• 5分钟
- Synthetic Data Augmentation• 2分钟
- Data-Driven Simulation• 3分钟
- Physics-Based Simulation• 2分钟
- Physics-Based Simulation Examples• 4分钟
1个作业• 总计15分钟
- Module 3 Quiz• 15分钟
In this module, you'll explore the concept of domain adaptation, a key aspect of transductive transfer learning. Domain adaptation helps you train models that perform well on a target domain, even when its data distribution differs from the source domain. You'll learn about the challenges of domain shift and labeled data scarcity and how these can impact model performance. We'll cover different types of domain adaptation, including unsupervised, semi-supervised, and supervised approaches. You'll also dive into techniques like Deep Domain Confusion (DDC), which integrates domain confusion loss into neural networks to create domain-invariant features. Additionally, you'll discover advanced methods such as Domain-Adversarial Neural Networks (DANNs), Correlation Alignment (CORAL), and Deep Adaptation Networks (DANs) that build on DDC to enhance domain adaptation by aligning feature distributions and capturing complex dependencies across network layers.
涵盖的内容
1个视频10篇阅读材料1个作业
1个视频• 总计6分钟
- Domain Adaptation• 6分钟
10篇阅读材料• 总计143分钟
- Domain Adaptation: Background• 1分钟
- Unsupervised, Semi-Supervised & Supervised• 10分钟
- Deep Domain Confusion• 8分钟
- Related Work Based on DDC• 2分钟
- Deep Domain Confusion Architecture• 10分钟
- Implementation & Architecture• 10分钟
- Mathematical Formulation• 5分钟
- An Example Dataset: Office-31• 2分钟
- An Example DDC Experiment• 5分钟
- Transfer Learning Practice Activity• 90分钟
1个作业• 总计10分钟
- Module 4 Quiz• 10分钟
In this module, we’ll explore weak supervision, a technique for training machine learning models with limited, noisy, or imprecise labels. You'll learn about different types of weak supervision and why they are crucial in small data domains. We’ll cover techniques such as semi-supervised learning, self-supervised learning, and active learning, along with advanced methods such as Temporal Ensembling and the Mean Teacher approach. Additionally, you'll discover Bayesian deep learning and active learning strategies to improve training efficiency. Finally, you'll see real-world applications in fields like medical imaging, NLP, fraud detection, autonomous driving, and biology.
涵盖的内容
1个视频8篇阅读材料1个作业
1个视频• 总计7分钟
- What is Weak Supervision?• 7分钟
8篇阅读材料• 总计54分钟
- Types of Weak Supervision• 6分钟
- Semi-Supervised Learning• 10分钟
- Self-Supervised Learning• 15分钟
- Active Learning• 6分钟
- Applications of Weak Supervision• 2分钟
- Case Study: Medical Imaging• 5分钟
- Case Study: Autonomous Driving• 5分钟
- Case Study: Natural Language Processing• 5分钟
1个作业• 总计30分钟
- Module 5 Quiz• 30分钟
In this module, you'll explore how Zero-Shot Learning (ZSL) enables models to recognize new categories without having seen any examples of those categories during training. This is achieved by leveraging intermediate semantic descriptions, such as attributes, shared between seen and unseen classes. You'll also learn about the importance of regularization in preventing overfitting and improving generalization, as well as how generative models like GANs and VAEs enhance ZSL by synthesizing unseen class data. Additionally, we'll examine Generalized Zero-Shot Learning (GZSL), which tests models on both seen and unseen classes, making the task more challenging and realistic. By the end of this module, you'll have a solid understanding of how ZSL and its extensions can be applied to various machine learning tasks.
涵盖的内容
1个视频9篇阅读材料1个作业
1个视频• 总计5分钟
- Generalized Zero-Shot Learning• 5分钟
9篇阅读材料• 总计71分钟
- Introduction to Zero-Shot Learning• 3分钟
- ZSL: Notation and Problem Setup• 3分钟
- Learning a Linear Predictor for Seen Classes• 10分钟
- Problem Extension for ZSL: From Seen to Unseen Classes• 15分钟
- An Embarrassingly Simple Approach to ZSL• 10分钟
- ZSL with Generative Models• 10分钟
- Generalized Zero-Shot Learning (GZSL)• 10分钟
- Zero-Shot Learning: Semantic Autoencoders• 5分钟
- Generalized ZSL With Generative Models• 5分钟
1个作业• 总计30分钟
- Module 6 Quiz• 30分钟
This module focuses on Few-Shot Learning (FSL), a critical paradigm in machine learning that enables models to classify new examples with only a small number of labeled instances. Unlike traditional deep learning models that require vast amounts of labeled data, FSL mimics the human ability to generalize from limited examples, making it highly useful for tasks like image classification, object detection, and natural language processing (NLP). The lecture introduces Matching Networks, a metric-based learning approach designed to solve one-shot learning problems by learning a similarity function that maps new examples to previously seen labeled instances. Students will gain an in-depth understanding of how nearest-neighbor approaches, differentiable embedding functions, and attention mechanisms help in optimizing few-shot learning models. Through discussions, theoretical formulations, and real-world applications, this lecture equips students with practical insights into how AI can function effectively in data-scarce environments.
涵盖的内容
1个视频7篇阅读材料1个作业
1个视频• 总计6分钟
- Introduction to Few-Shot Learning• 6分钟
7篇阅读材料• 总计46分钟
- What is Few-Shot Learning?• 10分钟
- Introduction to One-Shot Learning• 2分钟
- Matching Networks: An Approach to One-Shot Learning• 10分钟
- Training Matching Networks• 3分钟
- Improving Few-Shot Visual Classification• 10分钟
- Enhancing Few-Shot Image Classification With Unlabeled Examples• 10分钟
- Congratulations• 1分钟
1个作业• 总计30分钟
- Module 7 Quiz• 30分钟
位教师

提供方

提供方

Founded in 1898, Northeastern is a global research university with a distinctive, experience-driven approach to education and discovery. The university is a leader in experiential learning, powered by the world’s most far-reaching cooperative education program. The spirit of collaboration guides a use-inspired research enterprise focused on solving global challenges in health, security, and sustainability.
从 Machine Learning 浏览更多内容
NNortheastern University
课程
PPearson
课程
OO.P. Jindal Global University
课程
EEdge Impulse
课程
人们为什么选择 Coursera 来帮助自己实现职业发展

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.
常见问题
To access the course materials, assignments and to earn a Certificate, you will need to purchase the Certificate experience when you enroll in a course. You can try a Free Trial instead, or apply for Financial Aid. The course may offer 'Full Course, No Certificate' instead. This option lets you see all course materials, submit required assessments, and get a final grade. This also means that you will not be able to purchase a Certificate experience.
When you purchase a Certificate you get access to all course materials, including graded assignments. Upon completing the course, your electronic Certificate will be added to your Accomplishments page - from there, you can print your Certificate or add it to your LinkedIn profile.
Yes. In select learning programs, you can apply for financial aid or a scholarship if you can’t afford the enrollment fee. If fin aid or scholarship is available for your learning program selection, you’ll find a link to apply on the description page.
更多问题
提供助学金,

