This course addresses the challenge of machine learning (ML) in the context of small datasets, a significant issue due to ML's increasing data demands. Despite ML's success in various fields, many areas can't provide large labeled datasets because of costs, privacy, or security laws. As big data becomes standard, efficiently learning from smaller datasets is crucial. This course, ideal for graduate students with some ML experience, focuses on modern deep learning techniques for small data applications relevant in healthcare, military, and various industry sectors. Prerequisites include ML familiarity and Python proficiency. Deep learning experience is not necessary but beneficial.

Machine Learning with Small Data Part 1

Skills you'll gain
Tools you'll learn
Details to know

Add to your LinkedIn profile
8 assignments
See how employees at top companies are mastering in-demand skills

There are 7 modules in this course
In this module, we will explore the pivotal role of data as the foundation for machine learning algorithms. We begin by discussing the significance of large datasets in training deep learning models as these datasets are crucial for the models’ successful application and effectiveness. We will also delve into the challenges associated with small datasets, particularly in sensitive fields such as healthcare and defense, where data acquisition is often difficult, costly, or subject to stringent privacy and security regulations. To address these challenges, the course will introduce various strategies for making the most of limited data, including data-efficient machine learning techniques and the use of synthetic data augmentation. Additionally, we will present the course structure and discuss a curated selection of research papers that align with and enrich our course topics.
What's included
2 videos13 readings1 assignment
2 videos•Total 16 minutes
- Data Matters•8 minutes
- Setting Up Your Local Environment•8 minutes
13 readings•Total 81 minutes
- Course Overview•1 minute
- Syllabus - Machine Learning for Small Data•10 minutes
- Academic Integrity•1 minute
- Data Matters—Especially for Deep Learning•2 minutes
- Data-Parameters-Power Scaling in AI Model•5 minutes
- Exponential Growth of Training Data•10 minutes
- Exponential Growth of Model Complexity•5 minutes
- Exponential Growth in Computational Resources•5 minutes
- The Scale Paradox: When Smaller ML Models Outperform Giants•5 minutes
- Large Datasets for Deep Learning•10 minutes
- What is Small Data?•2 minutes
- Installing PyTorch•5 minutes
- Large vs. Small Datasets in Machine Learning•20 minutes
1 assignment•Total 10 minutes
- Module 1 Quiz•10 minutes
In this module, we will delve into the core aspects of machine learning with a focus on the importance of data, particularly in deep learning applications. We start by emphasizing how large datasets are essential for training deep learning models effectively, as they enable the models to capture and learn from complex patterns, improving their overall performance. Additionally, we'll explore the intersection of data availability, computational power, and model capacity, highlighting how these elements interact to refine model accuracy and efficiency. Furthermore, the module will cover computing advancements beyond Moore's Law and their impact on machine learning, illustrating how modern hardware like CPUs, GPUs, and TPUs enhance computational capabilities critical for training sophisticated models. We'll also delve into scaling laws in deep learning, discussing empirical findings that show how model performance improves predictably with increases in dataset size and model complexity, although with diminishing returns. To provide a deeper theoretical foundation, we'll examine the Vapnik-Chervonenkis (VC) theory, which offers insights into how learning curves and model complexity relate to a model’s ability to generalize from training data. This discussion will extend to practical applications and theoretical limitations, helping to frame machine learning challenges in terms of data sufficiency, model fitting, and the balance between bias and variance. By the end of this module, students will have a thorough understanding of the dynamic interplay between these factors and their implications for machine learning practice and research.
What's included
1 video19 readings2 assignments1 app item
1 video•Total 9 minutes
- Machine Learning Model Performance•9 minutes
19 readings•Total 144 minutes
- Ingredients Relationship•10 minutes
- Computing Power: Growth Beyond Moore’s Law•10 minutes
- Scaling Laws•5 minutes
- Learning Curves•15 minutes
- Model Capacity Required to Fit Data•3 minutes
- Model Performance and Dataset Size•2 minutes
- Model Performance and Model Capacity•2 minutes
- Bias-Variance Trade-Off•15 minutes
- From a Linear Algebra Perspective•2 minutes
- Underdetermined Problems and Overparameterized Models•8 minutes
- Revisiting Bias-Variance with Double Descent•8 minutes
- Comparison of Learning Paradigms•15 minutes
- A Learning Machine•2 minutes
- How Do We Characterize Model Complexity?•1 minute
- Vapnik–Chervonenkis (VC) Dimension - Shattering•10 minutes
- Notions of VC Dimension•10 minutes
- Examples of Shattering and VC Dimension•10 minutes
- VC Dimension in Neural Networks•15 minutes
- Resources•1 minute
2 assignments•Total 60 minutes
- Module 2 Quiz•30 minutes
- Calculating the VC Dimension of SVM Models•30 minutes
1 app item•Total 10 minutes
- Examples of Learning Machines•10 minutes
In this module, we’ll explore transfer learning and its role in data-efficient machine learning, where models leverage knowledge from previous tasks to improve performance on new, related tasks. We’ll also cover various types of transfer learning, including transductive, inductive, and unsupervised methods, each addressing different challenges and applications. We’ll discuss some practical steps for implementing transfer learning, such as selecting and fine-tuning pre-trained models, to reduce reliance on large datasets. We’ll also examine data-driven and physics-based simulations for data augmentation, highlighting their use in enhancing training under constrained conditions. Finally, we’ll review key papers on transfer learning techniques to address data scarcity and improve model performance.
What's included
1 video15 readings1 assignment
1 video•Total 6 minutes
- Transfer Learning•6 minutes
15 readings•Total 72 minutes
- Data-efficient Machine Learning•10 minutes
- Leveraging Pre-trained Models for Efficient Machine Learning•2 minutes
- Vanilla Transfer Learning •2 minutes
- Types of Transfer Learning•2 minutes
- Transductive Transfer Learning Algorithms•10 minutes
- Inductive Transfer Learning Algorithms•10 minutes
- Transductive Examples I•5 minutes
- Transductive Examples II•5 minutes
- Transductive Examples III•5 minutes
- Inductive Examples•5 minutes
- Multi-Task Learning & Meta-Learning•5 minutes
- Synthetic Data Augmentation•2 minutes
- Data-Driven Simulation•3 minutes
- Physics-Based Simulation•2 minutes
- Physics-Based Simulation Examples•4 minutes
1 assignment•Total 15 minutes
- Module 3 Quiz•15 minutes
In this module, you'll explore the concept of domain adaptation, a key aspect of transductive transfer learning. Domain adaptation helps you train models that perform well on a target domain, even when its data distribution differs from the source domain. You'll learn about the challenges of domain shift and labeled data scarcity and how these can impact model performance. We'll cover different types of domain adaptation, including unsupervised, semi-supervised, and supervised approaches. You'll also dive into techniques like Deep Domain Confusion (DDC), which integrates domain confusion loss into neural networks to create domain-invariant features. Additionally, you'll discover advanced methods such as Domain-Adversarial Neural Networks (DANNs), Correlation Alignment (CORAL), and Deep Adaptation Networks (DANs) that build on DDC to enhance domain adaptation by aligning feature distributions and capturing complex dependencies across network layers.
What's included
1 video10 readings1 assignment
1 video•Total 6 minutes
- Domain Adaptation•6 minutes
10 readings•Total 143 minutes
- Domain Adaptation: Background•1 minute
- Unsupervised, Semi-Supervised & Supervised•10 minutes
- Deep Domain Confusion•8 minutes
- Related Work Based on DDC•2 minutes
- Deep Domain Confusion Architecture•10 minutes
- Implementation & Architecture•10 minutes
- Mathematical Formulation•5 minutes
- An Example Dataset: Office-31•2 minutes
- An Example DDC Experiment•5 minutes
- Transfer Learning Practice Activity•90 minutes
1 assignment•Total 10 minutes
- Module 4 Quiz•10 minutes
In this module, we’ll explore weak supervision, a technique for training machine learning models with limited, noisy, or imprecise labels. You'll learn about different types of weak supervision and why they are crucial in small data domains. We’ll cover techniques such as semi-supervised learning, self-supervised learning, and active learning, along with advanced methods such as Temporal Ensembling and the Mean Teacher approach. Additionally, you'll discover Bayesian deep learning and active learning strategies to improve training efficiency. Finally, you'll see real-world applications in fields like medical imaging, NLP, fraud detection, autonomous driving, and biology.
What's included
1 video8 readings1 assignment
1 video•Total 7 minutes
- What is Weak Supervision?•7 minutes
8 readings•Total 54 minutes
- Types of Weak Supervision•6 minutes
- Semi-Supervised Learning•10 minutes
- Self-Supervised Learning•15 minutes
- Active Learning•6 minutes
- Applications of Weak Supervision•2 minutes
- Case Study: Medical Imaging•5 minutes
- Case Study: Autonomous Driving•5 minutes
- Case Study: Natural Language Processing•5 minutes
1 assignment•Total 30 minutes
- Module 5 Quiz•30 minutes
In this module, you'll explore how Zero-Shot Learning (ZSL) enables models to recognize new categories without having seen any examples of those categories during training. This is achieved by leveraging intermediate semantic descriptions, such as attributes, shared between seen and unseen classes. You'll also learn about the importance of regularization in preventing overfitting and improving generalization, as well as how generative models like GANs and VAEs enhance ZSL by synthesizing unseen class data. Additionally, we'll examine Generalized Zero-Shot Learning (GZSL), which tests models on both seen and unseen classes, making the task more challenging and realistic. By the end of this module, you'll have a solid understanding of how ZSL and its extensions can be applied to various machine learning tasks.
What's included
1 video9 readings1 assignment
1 video•Total 5 minutes
- Generalized Zero-Shot Learning•5 minutes
9 readings•Total 71 minutes
- Introduction to Zero-Shot Learning•3 minutes
- ZSL: Notation and Problem Setup•3 minutes
- Learning a Linear Predictor for Seen Classes•10 minutes
- Problem Extension for ZSL: From Seen to Unseen Classes•15 minutes
- An Embarrassingly Simple Approach to ZSL•10 minutes
- ZSL with Generative Models•10 minutes
- Generalized Zero-Shot Learning (GZSL)•10 minutes
- Zero-Shot Learning: Semantic Autoencoders•5 minutes
- Generalized ZSL With Generative Models•5 minutes
1 assignment•Total 30 minutes
- Module 6 Quiz•30 minutes
This module focuses on Few-Shot Learning (FSL), a critical paradigm in machine learning that enables models to classify new examples with only a small number of labeled instances. Unlike traditional deep learning models that require vast amounts of labeled data, FSL mimics the human ability to generalize from limited examples, making it highly useful for tasks like image classification, object detection, and natural language processing (NLP). The lecture introduces Matching Networks, a metric-based learning approach designed to solve one-shot learning problems by learning a similarity function that maps new examples to previously seen labeled instances. Students will gain an in-depth understanding of how nearest-neighbor approaches, differentiable embedding functions, and attention mechanisms help in optimizing few-shot learning models. Through discussions, theoretical formulations, and real-world applications, this lecture equips students with practical insights into how AI can function effectively in data-scarce environments.
What's included
1 video7 readings1 assignment
1 video•Total 6 minutes
- Introduction to Few-Shot Learning•6 minutes
7 readings•Total 46 minutes
- What is Few-Shot Learning?•10 minutes
- Introduction to One-Shot Learning•2 minutes
- Matching Networks: An Approach to One-Shot Learning•10 minutes
- Training Matching Networks•3 minutes
- Improving Few-Shot Visual Classification•10 minutes
- Enhancing Few-Shot Image Classification With Unlabeled Examples•10 minutes
- Congratulations•1 minute
1 assignment•Total 30 minutes
- Module 7 Quiz•30 minutes
Instructor

Offered by

Offered by

Founded in 1898, Northeastern is a global research university with a distinctive, experience-driven approach to education and discovery. The university is a leader in experiential learning, powered by the world’s most far-reaching cooperative education program. The spirit of collaboration guides a use-inspired research enterprise focused on solving global challenges in health, security, and sustainability.
Why people choose Coursera for their career

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.
Explore more from Data Science
NNortheastern University
Course
PPearson
Course
OO.P. Jindal Global University
Course
EEdge Impulse
Course