Empowering GenAI on AWS EKS: Deploy, Experiment, and Scale Advanced AI Workflows

2 Weeks

·

Cohort-based Course

Unlock the power of AI with hands-on expertise in vLLM, JARK Satck, RayServe, BioNeMo, Kubernetes, Amazon EKS, and Terraform

Previously at

Spryker Systems
ING
Deutsche Bank
Renault Group
Autohero Deutschland

Course overview

Deploy, Train & Scale GenAI on AWS EKS: Master Production-Grade AI

Unlock the power of AI with hands-on expertise in Kubernetes, Amazon EKS, Terraform, vLLM, JARK Satck, RayServe & BioNeMo.


Why This Course?


The ability to build, deploy, and scale AI applications efficiently is no longer a luxury—it’s a necessity.


AI engineers, DevOps professionals, and software developers who can bridge the gap between ML and scalable infrastructure are in high demand.


This course will equip you with end-to-end, production-grade AI deployment skills using Amazon EKS, Kubernetes, Terraform, JARK Stack (Jypyter, Argo Workflows, Ray, Kubernetes), GPU, vLLM, Grafana, Prometheus and BioNeMo, and more—the same technologies used by leading AI-driven organizations.


You'll learn not just theory, but practical, real-world workflows for LLM deployment, distributed training, GPU optimization, and MLOps automation—ensuring you're ready to build and scale AI solutions at any organization.


🎯 Course Milestones:


You'll begin by mastering the fundamentals of Generative AI and ML workflows, and understanding how Kubernetes enables large-scale model deployment, focusing on Amazon EKS for enterprise-ready solutions.


Next, you’ll provision GPU-enabled EKS clusters, configure auto-scaling using Karpenter and Terraform, and optimize networking and IAM roles, ensuring a robust and automated infrastructure for AI workloads.


With the foundation set, you'll dive into the JARK stack (JupyterHub, Argo Workflows, Ray, and Kubernetes) for seamless model experimentation, workflow automation, and scaling distributed workloads.


From experimentation, you’ll transition into scalable inference, leveraging RayServe and vLLM to deploy high-performance LLMs, optimize GPU memory, and integrate Open WebUI for real-time interactions.


To complete the ML and GenAI pipeline, you'll tackle distributed model training with BioNeMo, handling large datasets, multi-GPU orchestration, and performance tuning for efficient large-scale learning.


By the final stage, you’ll be ready for your graduation project—building a real-world GenAI-powered chat application, integrating the entire ML workflow from training to inference while implementing CI/CD and monitoring strategies.


Through hands-on labs and project-based learning, you'll gain the practical expertise to deploy enterprise AI solutions, advance your career, and step into leadership roles in cloud-native AI engineering.


🌻What You’ll Gain


This course transforms your career by giving you the hands-on experience and expertise needed to operate at the intersection of AI, ML, DevOps, and cloud-native architectures.


🚀 By the end of this course, you will:


Deploy AI Models at Scale – Learn Kubernetes and Amazon EKS to run AI workloads on GPU-powered clusters with high availability and efficiency.


Master AI Inference with RayServe – Scale your LLM deployments seamlessly with RayServe and vLLM, ensuring fast and cost-effective inferencing.


Optimize AI Workloads – Use BioNeMo and distributed training techniques to maximize GPU utilization and optimize ML model training at scale.


Hands-on Real-World Projects – Work on a full-fledged GenAI chat application, gaining practical expertise that applies directly to industry use cases.


Career Acceleration – Position yourself as an AI/ML platform engineer, MLOps expert, or cloud-native AI specialist, opening doors to high-paying roles in top-tier companies.


🎓 Why Learn from This Course?


Most AI courses focus on model training but neglect the critical aspects of deploying, scaling, and operationalizing AI systems. This course bridges the gap.


You won’t just learn how to train AI models—you’ll learn how to deploy, optimize, and maintain them at scale in real-world environments.


✔️ Designed for professionals who want to apply DevOps & Kubernetes to ML


✔️ Project-based learning ensures practical expertise


✔️ Covers the entire AI lifecycle from training to production deployment


✔️ Instructor-led deep dives on critical AI deployment patterns


This is the only course that provides a complete skill set in AI, ML, DevOps, and cloud-native architectures—ensuring you stay ahead of the curve in one of the most sought-after fields in tech today.

DevOps, AI & ML, Software Engineers, and Entry-Level Engineers, this course is for you!

01

DevOps professionals looking to apply their cloud, Kubernetes, and automation skills to AI/ML workloads on AWS.

02

AI/ML engineers who want to learn Kubernetes, cloud automation, and scalable model deployment using DevOps best practices.

03

Software Engineers Expanding into AI, ML & MLOps: Software developers who want to gain expertise in AI/ML model training, deployment, and ML

Mastering AI Engineering Using Cloud-Native Technologies and DevOps Best Practices

Setting Up Strong Foundations

Mastering the fundamentals of Generative AI and ML workflows, such as model training and model inference, while understanding the technologies that enable large-scale model deployment, focusing on Amazon Elastic Kubernetes Service for enterprise readiness.solutions.

Provision and configure GPU-enabled AWS EKS clusters using Terraform, with security, IAM, and networking best practices.

Give students an idea of how they can expect to grow throughout your course. Include specificity and precise results so students can benchmark exactly what they’ll learn.

Master Kubernetes basics: pods, deployments, autoscaling, resource limits, and troubleshooting logs for ML workloads.

Give students an idea of how they can expect to grow throughout your course. Include specificity and precise results so students can benchmark exactly what they’ll learn.

Deploy the JARK stack with JupyterHub, Argo Workflows, and Ray to streamline collaborative ML experiments.

Give students an idea of how they can expect to grow throughout your course. Include specificity and precise results so students can benchmark exactly what they’ll learn.

Launch inference endpoints using Docker, RayServe, vLLM, and HuggingFace libraries on AWS EKS for AI chat apps.

Give students an idea of how they can expect to grow throughout your course. Include specificity and precise results so students can benchmark exactly what they’ll learn.

Analyze GPU memory usage and optimize model inference performance using real-time metrics and port-forwarding.

Execute distributed training with BioNeMo and Kubeflow, leveraging multi-GPU setups and FSx for Lustre storage.

Learn Kubernetes and AWS EKS security and operation excellence best practices.

Build a complete GenAI chat app from experimentation to training and inference with hands-on labs and milestones.

Utilize the standard DevOps and Cloud Native tools to build AI/ML platforms, including Terraform, Kubernetes, Grafana, Karpenter, and Helm.

What’s included

Aymen Segni

Live sessions

Learn directly from Aymen Segni in a real-time, interactive format.

Lifetime access

Go back to course content and recordings whenever you need to.

Community of peers

Stay accountable and share insights with like-minded professionals.

Certificate of completion

Share your new skills with your employer or on LinkedIn.

Maven Guarantee

This course is backed by the Maven Guarantee. Students are eligible for a full refund up until the halfway point of the course.

Course syllabus

8 live sessions • 7 lessons • 7 projects

Week 1

Sep 1—Sep 7

    📅 Week 1: April 2 – April 6 | Class I – Foundations & EKS Deployment

    0 items

    Module I.1 –Introduction to GenIA, ML and Kubernetes

    • Apr

      3

      📌 Live Event I.1: April 3 – Foundations of GenAI, ML & Kubernetes

      Thu 4/34:00 PM—5:15 PM (UTC)
    2 more items

    📌 Module I.2 – Deploying a Production-Ready AWS EKS Cluster

    • Apr

      4

      📌 Live Event: April 4 – Hands-On Lab: Deploy AWS EKS with GPU Support

      Fri 4/410:00 AM—11:30 AM (UTC)
    2 more items

Week 2

Sep 8—Sep 14

    📅 Week 2: April 7 – April 13 | Class II – Model Experimentation & Inference

    0 items

    Module II.1 – GenAI and LLM Models Experiments with the JRAK Stack

    • Apr

      8

      📌 Live Event: April 8 – Hands-On Lab: LLM Experimentation on AWS EKS

      Tue 4/84:00 PM—5:30 PM (UTC)
    2 more items

    Module II.1 – Model Inference Using RayServe and vLLM

    • Apr

      10

      📌 Live Event: April 10 – Hands-On Lab: Deploying an LLM Inference Endpoint

      Thu 4/104:00 PM—5:30 PM (UTC)
    2 more items

    Module II.3 – Distributed Model Training with BioNeMo

    • Apr

      12

      📌 Live Event: April 12 – Advanced AI Training with BioNeMo

      Sat 4/121:00 AM—2:30 AM (UTC)
    2 more items

Week 3

Sep 15

    Week 3: April 14 – April 16 | Class III – Graduation Project

    • Apr

      16

      📌 Live Event: April 16 – Graduation Project Showcase & Q&A

      Wed 4/164:00 PM—5:30 PM (UTC)
    2 more items

    Graduation Project – Building a Real-World GenAI Application

    2 items

    Apr

    16

    Graduation Project: Develop a complete GenAI chat application

    Wed 4/169:00 PM—10:15 PM (UTC)

Bonus

    🔥 Bonus Module: Deploying DeepSeek R1 on AWS EKS

    • Feb

      18

      🔥 Live Event Bonus Module - Apr. 18: Deploying DeepSeek R1 on AWS EKS

      Tue 2/185:00 PM—6:30 PM (UTC)

Meet your instructor

Aymen Segni

Aymen Segni

Cloud & DevOps Expert specializing in modern infrastructure with Kubernetes

Aymen Segni is a Cloud & DevOps Expert, with deep expertise in Cloud Infrastructure, SRE, Data Platforms, Kubernetes, and Cloud-Native architectures. With a background in software engineering and systems architecture,


Aymen has helped startups and enterprises design, deploy, and scale their infrastructure and applications on platforms like AWS and Kubernetes.

His career spans Cloud Platform engineering, DevOps, and AI/ML operations, making him uniquely positioned to bridge the gap between AI model development and scalable, production-grade deployments. Aymen has worked with cloud providers, FineTech industry, software vendors, e-commerce, and consulting, leading teams to optimize technologies to make the business run better.


As a mentor and educator, Aymen has trained engineers in Kubernetes, cloud automation, and AI/ML deployment strategies, helping professionals transition into AI-driven DevOps roles. This course distills years of hands-on experience into a practical, project-based learning journey, ensuring students gain real-world skills for high-impact AI engineering careers.

A pattern of wavy dots

Be the first to know about upcoming cohorts

Empowering GenAI on AWS EKS: Deploy, Experiment, and Scale Advanced AI Workflows

Free resource

DeepSeek-R1 on AWS Kubernetes Service EKS with Ray and vLLM

This course meets the need for scalable, enterprise-grade AI. As generative AI evolves, organizations require models that are powerful, resource-efficient, and resilient. Master AWS EKS, Ray, and vLLM to build applications that deliver real-time insights, drive innovation, and create transformative business impact.

Get this free resource

Learning is better with cohorts

Learning is better with cohorts

Active hands-on learning

This course builds on live workshops and hands-on projects

Interactive and project-based

You’ll be interacting with other learners through breakout rooms and project teams

Learn with a cohort of peers

Join a community of like-minded people who want to learn and grow alongside you

Frequently Asked Questions

A pattern of wavy dots

Be the first to know about upcoming cohorts

Empowering GenAI on AWS EKS: Deploy, Experiment, and Scale Advanced AI Workflows