By completing this course, you'll master building powerful machine learning systems that excel with limited data. You'll gain expertise in multi-task learning, meta-learning, and advanced data augmentation—from physics-based simulations to generative approaches—enabling models to adapt quickly and perform beyond their dataset size.



Skills you'll gain
Details to know

Add to your LinkedIn profile
September 2025
7 assignments
See how employees at top companies are mastering in-demand skills

There are 7 modules in this course
In this module, we will introduce the fundamentals of Multi-Task Learning (MTL), a paradigm where multiple related tasks are learned simultaneously by sharing representations. This approach leverages the commonalities among tasks to improve generalization, reduce overfitting, and achieve better performance with fewer training examples. We will explore how MTL is applied across various domains, such as natural language processing, computer vision, and speech recognition, and examine practical examples such as using MTL to enhance image classification and object detection in autonomous systems. Students will gain insights into both the benefits and challenges of MTL, including issues such as task imbalance, negative transfer, and scalability. Additionally, we will delve into meta-learning techniques, such as Conditional Neural Adaptive Processes (CNAPs), that extend MTL by enabling models to quickly adapt to new tasks with minimal data.
What's included
1 video15 readings1 assignment
This module explores the concept of meta-learning, or "learning to learn," which enables models to generalize across various tasks by leveraging knowledge from similar tasks. We will delve into key meta-learning algorithms such as Model-Agnostic Meta-Learning (MAML) and Prototypical Networks and examine their applications in computer vision using datasets such as ImageNet, Omniglot, CUB-200-2011, and FGVC-Aircraft. The module also covers the Meta-Dataset framework, which provides a diverse range of tasks for training robust and adaptable meta-learning models.
What's included
1 video7 readings1 assignment
This module focuses on generative models for data augmentation, covering key generative AI techniques that enhance machine learning applications by generating synthetic but realistic data. We begin by introducing generative adversarial networks (GANs), Variational Autoencoders (VAEs), Normalizing Flows, Diffusion Models, and Motion Graphs, highlighting their mathematical foundations, training mechanisms, and real-world applications. Additionally, we discuss the limitations of each model and the computational challenges they present. The lecture provides insights into how generative models contribute to modern AI systems, including image synthesis, domain adaptation, super-resolution, motion synthesis, and data augmentation in small-data learning scenarios.
What's included
1 video28 readings1 assignment
This module focuses on physics-based simulation for data augmentation, exploring how physics-driven techniques generate realistic synthetic data to enhance machine learning models. We will discuss key advantages of physics-based simulations, such as scalability, cost-effectiveness, and their ability to model rare events. The module also covers notable approaches, including GeoNet (CVPR 2018) for depth and motion estimation, ScanAva (ECCVW 2018) for semi-supervised learning with 3D avatars, and SMPL (ACM Transactions on Graphics, Volume 15) for human body modeling. Additionally, we introduce equation-based simulation techniques such as Finite Element Method (FEM) and Navier-Stokes equations for modeling fluid dynamics. The module highlights challenges in bridging the simulation-to-reality gap and optimizing computational costs while ensuring high-fidelity synthetic data generation.
What's included
1 video10 readings1 assignment
This module introduces Neural Radiance Fields (NeRF), a deep learning-based approach for synthesizing novel views of complex 3D scenes. Unlike traditional 3D reconstruction techniques such as Structure-from-Motion (SfM) and Multi-View Stereo (MVS), which rely on explicit point cloud representations, NeRF learns a continuous volumetric representation of a scene using a fully connected neural network. By taking a set of 2D images captured from different viewpoints, NeRF estimates the density and color of light rays at each spatial location, enabling high-quality, photorealistic novel view synthesis. The lecture also explores how NeRF improves upon prior methods, such as depth estimation, photogrammetry, and classic geometric techniques. Understanding NeRF provides valuable insights into data-efficient 3D scene representation—a critical area for applications in computer vision, robotics, virtual reality (VR), and augmented reality (AR).
What's included
1 video6 readings1 assignment
This module explores diffusion models, a class of generative models that incrementally add noise to data and then learn to reverse the process to reconstruct high-quality samples. Diffusion models have gained prominence due to their state-of-the-art performance in image, video, and text generation, surpassing GANs in terms of sample quality and diversity. The module covers the foundational principles of Denoising Diffusion Probabilistic Models (DDPMs) and their training objectives, advancements such as Score-Based Generative Models, Latent Diffusion Models (LDMs), and Classifier-Free Guidance techniques. We also examine their real-world applications in text-to-image generation (Stable Diffusion, DALL·E), video synthesis (Sora, Veo 2), and high-resolution image synthesis. Finally, the module provides insights into the mathematical framework, the optimization strategies, and the growing role of diffusion models in AI-driven content creation.
What's included
1 video11 readings1 assignment
This lecture explores 3D Gaussian Splatting (3DGS), a novel approach in computer vision for high-fidelity, real-time 3D scene rendering. Unlike traditional methods like Neural Radiance Fields (NeRF), which rely on continuous neural fields, 3DGS represents scenes using a collection of discrete anisotropic Gaussian functions. These Gaussians efficiently approximate scene geometry, radiance, and depth, enabling real-time rendering with minimal computational overhead. We discuss the theoretical foundations, mathematical formulations, and rendering techniques that make 3D Gaussian Splatting a game-changer in virtual reality (VR), augmented reality (AR), and interactive media. Additionally, we highlight key differences between isotropic and anisotropic Gaussian splats, their impact on rendering quality, and how optimization techniques refine their accuracy. Finally, we compare 3DGS to NeRF, analyzing their trade-offs in rendering speed, computational efficiency, and application suitability.
What's included
1 video6 readings1 assignment
Instructor

Offered by
Why people choose Coursera for their career





Open new doors with Coursera Plus
Unlimited access to 10,000+ world-class courses, hands-on projects, and job-ready certificate programs - all included in your subscription
Advance your career with an online degree
Earn a degree from world-class universities - 100% online
Join over 3,400 global companies that choose Coursera for Business
Upskill your employees to excel in the digital economy
Frequently asked questions
To access the course materials, assignments and to earn a Certificate, you will need to purchase the Certificate experience when you enroll in a course. You can try a Free Trial instead, or apply for Financial Aid. The course may offer 'Full Course, No Certificate' instead. This option lets you see all course materials, submit required assessments, and get a final grade. This also means that you will not be able to purchase a Certificate experience.
When you purchase a Certificate you get access to all course materials, including graded assignments. Upon completing the course, your electronic Certificate will be added to your Accomplishments page - from there, you can print your Certificate or add it to your LinkedIn profile.
You will be eligible for a full refund until two weeks after your payment date, or (for courses that have just launched) until two weeks after the first session of the course begins, whichever is later. You cannot receive a refund once you’ve earned a Course Certificate, even if you complete the course within the two-week refund period. See our full refund policy.
More questions
Financial aid available,