Pragmatic AI Labs

AI Orchestration: From local models to cloud

Pragmatic AI Labs

AI Orchestration: From local models to cloud

This course is part of AI Tooling Specialization

Alfredo Deza
Noah Gift

Instructors: Alfredo Deza

Access provided by Xavier School of Management, XLRI

Gain insight into a topic and learn the fundamentals.
Beginner level

Recommended experience

5 hours to complete
Flexible schedule
Learn at your own pace
Gain insight into a topic and learn the fundamentals.
Beginner level

Recommended experience

5 hours to complete
Flexible schedule
Learn at your own pace

What you'll learn

  • Build a prompt engineering pyramid from basic prompts to chain-of-thought reasoning in Rust, and evaluate decision factors for local vs cloud

  • Set up local AI infrastructure with Ollama, llamafile, aprender and Rust Candle GPU compilation, plus caching and RAG optimization strategies

  • Configure a production AI workstation with tmux, nvidia-smi, and Zenith, and integrate cloud workflows with AWS Spot, Hugging Face, and GitHub AI

Details to know

Shareable certificate

Add to your LinkedIn profile

Assessments

4 assignments

Taught in English
Recently updated!

April 2026

See how employees at top companies are mastering in-demand skills

 logos of Petrobras, TATA, Danone, Capgemini, P&G and L'Oreal

Build your subject-matter expertise

This course is part of the AI Tooling Specialization
When you enroll in this course, you'll also be enrolled in this Specialization.
  • Learn new concepts from industry experts
  • Gain a foundational understanding of a subject or tool
  • Develop job-relevant skills with hands-on projects
  • Earn a shareable career certificate

There are 4 modules in this course

A comprehensive course covering prompt engineering with chain-of-thought reasoning, local inference runtimes (Ollama, llamafile, Candle), GPU workstation configuration, and cost-optimized cloud deployment with AWS Spot instances.

What's included

7 videos3 readings1 assignment

Covers local vs cloud model tradeoffs, caching strategies, local RAG optimization, Ollama with custom Modelfiles, llamafile portable deployment, and Candle GPU-accelerated Rust inference.

What's included

9 videos3 readings1 assignment

Covers tmux session management, nvidia-smi and Zenith GPU monitoring, local workstation orchestration, AWS Spot instance deployment, Hugging Face and GitHub AI model workflows, and Rust project structure.

What's included

11 videos3 readings1 assignment

Head-to-head comparison of Ollama vs `apr` ([paiml/aprender](https://github.com/paiml/aprender)) running Qwen2.5-Coder-1.5B on the same prompt suite, same hardware. Build a chain-of-thought routing engine that selects runtimes based on task complexity and validation requirements, with cost analysis spanning local workstations, Spot instances, and Bedrock.

What's included

2 readings1 assignment

Earn a career certificate

Add this credential to your LinkedIn profile, resume, or CV. Share it on social media and in your performance review.

Instructors

Alfredo Deza
Pragmatic AI Labs
19 Courses633 learners

Offered by

Why people choose Coursera for their career

Felipe M.

Learner since 2018
"To be able to take courses at my own pace and rhythm has been an amazing experience. I can learn whenever it fits my schedule and mood."

Jennifer J.

Learner since 2020
"I directly applied the concepts and skills I learned from my courses to an exciting new project at work."

Larry W.

Learner since 2021
"When I need courses on topics that my university doesn't offer, Coursera is one of the best places to go."

Chaitanya A.

"Learning isn't just about being better at your job: it's so much more than that. Coursera allows me to learn without limits."

Explore more from Computer Science