Build and deploy a production serverless multi-model Artificial Intelligence (AI) system on Amazon Web Services (AWS) that integrates Amazon Bedrock and Ollama for cloud and local Large Language Model (LLM) execution. This capstone course, the final course in the Applied AI Engineering specialization, synthesizes 19 courses of prior learning into a comprehensive engineering project. You will implement Rust-based LLM applications using the Cargo Lambda toolchain for serverless deployment on AWS Lambda, design Yet Another Markup Language (YAML)-driven prompt engineering workflows for structured configuration management, and build multi-model flow orchestration that routes requests to appropriate models based on task requirements. The course begins with multi-model architecture fundamentals covering the evolving AI model ecosystem, model selection criteria for production workloads, and multi-provider integration patterns that enable fallback and cost optimization. You then advance to serverless production deployment, implementing an Amazon Bedrock router for dynamic model selection and deploying Rust serverless functions with Cargo Lambda that offer cold start and memory advantages for AI workloads. The final capstone challenge requires you to integrate multi-model orchestration, YAML prompt configuration, and serverless deployment into a complete production system evaluated against performance, cost, and reliability standards.

AI Tooling Capstone: Serverless Multi-Model Systems

AI Tooling Capstone: Serverless Multi-Model Systems
This course is part of AI Tooling Specialization


Instructors: Alfredo Deza
Access provided by KBTG
Gain insight into a topic and learn the fundamentals.
Beginner level
Recommended experience
4 hours to complete
Flexible schedule
Learn at your own pace
What you'll learn
Apply integration patterns using Amazon Bedrock for local and cloud-hosted model access, with performing LLM applications using Rust
Design prompt engineering workflows and multi flow orchestration routing to specialized models based on tasks, constraints, and performance
Deploy a serverless AI system on AWS Lambda, integrating Amazon Bedrock, prompt configuration, and reliable end-to-end production evaluation
Skills you'll gain
Details to know

Shareable certificate
Add to your LinkedIn profile
Assessments
3 assignments
Taught in English
Recently updated!
April 2026
See how employees at top companies are mastering in-demand skills

Build your subject-matter expertise
This course is part of the AI Tooling Specialization
When you enroll in this course, you'll also be enrolled in this Specialization.
- Learn new concepts from industry experts
- Gain a foundational understanding of a subject or tool
- Develop job-relevant skills with hands-on projects
- Earn a shareable career certificate

There are 3 modules in this course
Earn a career certificate
Add this credential to your LinkedIn profile, resume, or CV. Share it on social media and in your performance review.
Offered by
Why people choose Coursera for their career

Felipe M.
Learner since 2018
"To be able to take courses at my own pace and rhythm has been an amazing experience. I can learn whenever it fits my schedule and mood."

Jennifer J.
Learner since 2020
"I directly applied the concepts and skills I learned from my courses to an exciting new project at work."

Larry W.
Learner since 2021
"When I need courses on topics that my university doesn't offer, Coursera is one of the best places to go."

Chaitanya A.
"Learning isn't just about being better at your job: it's so much more than that. Coursera allows me to learn without limits."
Explore more from Computer Science

Pragmatic AI Labs

Pragmatic AI Labs

Pragmatic AI Labs


