Staff machine learning engineer, currently working as an AI consultant
7 people enrolled last week.
Transform your retrieval from “good enough” to “mission-critical” in weeks, not months. Most RAG systems stall in prototype purgatory: they demo well, but fail on complex queries—eroding trust and wasting engineering time. The difference isn’t just better tech, but a systematic mindset.
With the RAG Flywheel, you’ll:
✅ Pinpoint failures with synthetic evals
✅ Fine-tune embeddings for 20–40% gains
✅ Collect 5x more user feedback
✅ Segment queries to target high-impact fixes
✅ Build multimodal indices for docs, tables, images
✅ Route queries to the best retriever automatically
Week by week, you move from vague “make it better” to clear metrics, focused improvements, and compounding value. Real-world results include +20% accuracy from re-ranking, +14% with cross-encoders, and $50M revenue boosts from better search.
Join 400+ engineers applying this framework in production. Instructor Jason Liu has built multimodal retrieval and recommendation systems at Facebook, Stitch Fix, and through consulting—experience that shaped this practical, battle-tested approach.
Follow a repeatable process to continually evaluate and improve your RAG application
Evaluate retrieval quality using precision, recall, and MRR metrics to identify system weaknesses
Differentiate between leading metrics (experiments run) and lagging metrics (customer satisfaction) to drive actionable improvements
Design synthetic data generation pipelines that enable rapid experimentation without waiting for user data
Create comprehensive evaluation datasets using LLMs to generate realistic query-answer pairs
Establish baselines using tools like LanceDB to benchmark different retrieval implementations
Develop multimodal retrieval systems that handle documents, images, tables, and structured data
Synthesize lexical (BM25), semantic (embeddings), and metadata-based search for optimal results
Extract structured information from diverse data sources to enable precise filtering
Classify queries using domain expertise and few-shot classifiers to improve routing accuracy
A product leader, engineer, or data scientist looking to move beyond ad-hoc RAG prototypes into scalable, production-grade AI solutions.
A professional who understands LLM basics but wants a repeatable, data-driven methodology to improve retrieval relevance, latency, and user
Eager to create feedback loops that continuously refine and enhance the quality of RAG applications as models, data, and user needs evolve.
The goal of this course is not just to share with you a how-to guide, but rather how to systematically improve these architectures.
We have over 20 iPython notebooks that you can explore, run code to be more hands-on with the experiments that we plan to run.
Live sessions
Learn directly from Jason Liu in a real-time, interactive format.
6 Prerecorded Lectures
Short, focused videos that unpack the full RAG-improvement framework that you can rewatch anytime.
6+ Office Hour Q&As
Open office hours for deep dives, debugging help, and personalized feedback.
12 Hands-On Python Notebooks
Ready-to-run notebooks & walkthrough videos so you can practice every concept instantly.
Lifetime Slack Community
Private Slack for peer reviews, job leads, and ongoing support forever.
Expert Speaker Library
Curated talks from builders running large-scale RAG systems in production.
$2K+ in Cloud & AI Credits
Test vector DBs, LLM APIs, and infra with over $2,000 in partner credits.
Free Future Re-Enrollment
Join any future cohort at no cost and get updated content and live coaching again whenever you need it.
Certificate of completion
Showcase your advanced RAG skills to clients, employers, and your LinkedIn network.
Maven Guarantee
This course is backed by the Maven Guarantee. Students are eligible for a full refund up until the halfway point of the course.
23 live sessions • 41 lessons
Sep
16
Introductions
Sep
16
Office Hour
Sep
18
Optional: Watch Lecture
Sep
18
Office Hour
Sep
17
Context Rot: How Input Length Impacts LLM Performance [Kelly Hong]
Sep
23
Office Hour
Sep
25
Optional: Watch Lecture
Sep
25
Office Hour
Sep
24
Cheating at Query Understanding with LLMs [Doug Turnbull]
Learn to identify subtle failure modes where retrieval systems pass tests but disappoint users in production.
Master practical techniques to detect hallucinations and relevance issues before they impact end users.
Gain concrete architectural patterns to transform struggling RAG systems into reliable production applications.
Office hours: 1 hour per week
Pre-recorded lectures : 1 hour per week
Optional guest lectures. : 1-2 hours per week
Core sessions
Optional sessions
Sam Flamini
Nico Neven