In this course, you’ll take a comprehensive journey through the storage solutions available on Google Cloud, specifically tailored for AI and high-performance computing (HPC) workloads. You’ll learn how to choose the right storage for each stage of the ML lifecycle. You’ll explore how to optimize for I/O performance during training, manage massive datasets for data preparation, and serve model artifacts with low latency. Through practical examples and demonstrations, you’ll gain the expertise to design robust storage solutions that accelerate your AI innovation.

AI Infrastructure: Storage Options

AI Infrastructure: Storage Options

Instructor: Google Cloud Training
Access provided by Masterflex LLC, Part of Avantor
What you'll learn
Determine the appropriate storage options and storage best practices for each phase of the AI data pipeline.
Determine the right storage solutions within each phase of the AI data pipeline.
Identify storage options and techniques for data preparation, model training, model serving, and data archiving.
Explore example storage architectures for model training and serving.
Skills you'll gain
Details to know

Add to your LinkedIn profile
3 assignments
December 2025
See how employees at top companies are mastering in-demand skills

There are 5 modules in this course
This module offers an overview of the course and outlines the learning objectives.
What's included
1 plugin
This module details the role of storage infrastructure in the AI data pipeline. It covers performance demands, key Google Cloud solutions, and the decision criteria for selecting a service based on capacity, throughput, and latency.
What's included
1 assignment2 plugins
This module details the critical phases of data preparation and model training within the AI workflow. It covers optimizing data loading using Cloud Storage, Anywhere Cache, and the Dataflux Dataset tool, while comparing high-performance file systems like Cloud Storage FUSE and Managed Lustre. Additionally, it outlines decision criteria for efficient checkpointing strategies to ensure fault tolerance and minimize GPU idle time.
What's included
1 assignment5 plugins
This module details strategies for AI model serving and data archiving. It covers selecting storage—Managed Lustre, Cloud Storage, or Hyperdisk ML—based on scale and latency, and optimization techniques, like GKE Image Streaming and Cloud Storage FUSE, to minimize costs and load times.
What's included
1 assignment6 plugins
Student PDF links to all modules
What's included
1 reading
Instructor

Offered by
Why people choose Coursera for their career

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.
Explore more from Information Technology

Google Cloud

Google Cloud



