10 people enrolled last week.
Keyword search + an agent - is that all we need now?
The agentic world challenges traditional ways of building search. Traditionally we've built thick, complex search APIs - query understanding, rerankers, learning to rank, vector databases. Yet an agent can compete in deep research benchmarks with the dumbest of search tools.
This course pushes you to rethink search. We teach:
Replacing search with agents - A search agent learns about retrieval systems, tailoring searches, achieving competitive results
Guiding agents, not algorithms - Search developers won't train algorithms. They'll build harnesses
Steering with an LLM judge - Integrate human and clickstream feedback to guide the agent's judgment towards what users want
Autoresearching to smarter search - Use + build a coding agent that optimizes retrieval code like an ML model.
Come learn alongside search industry expert Doug Turnbull as we unpack how agents will rewrite the search industry.
Rethink search to put tireless agents at work for human users
Build search backends agents can reason about
Why agents love simplicity over complexity
Focusing tools on domain tasks, not "search"
Sidestep complex NLP to better understand queries
Iterate on query understanding through agentic approaches
Organize and structure content to be retrievable by an agent
How to integrate agentic approaches into a traditional search stack
Best practices to manage the agentic loop to save tokens, money, and prevent context rot
Technical approaches to extracting agentic insights (ie code generation)
An agent will by default use its judgments to iterate on relevance on its own. But shouldn't it use humans?
Integrating external feedback (user clicks, etc) into agentic relevance
Beyond LLM as a Judge to exploring how to iterate and improve retrieval for agents and humans
Search technologists that want to learn how to ensure RAG and agentic search win
Startups eager to get a fast start with a modern approach to search, sidestepping old, outdated practices
Data scientists that want to consider how agentic / LLM / RAG based search can be oriented towards user feedback and business outcomes
We will build on top of existing, core search knowledge
We will use common search datasets that use metrics like NDCG, etc to measure search
Vector search will be one tool in our toolkit we'll apply to the agentic tools that we build

Live sessions
Learn directly from Doug Turnbull in a real-time, interactive format.
Lifetime access
Go back to course content and recordings whenever you need to.
Community of peers
Stay accountable and share insights with like-minded professionals.
Certificate of completion
Share your new skills with your employer or on LinkedIn.
Maven Guarantee
Your purchase is backed by the Maven Guarantee.
May
18
May
19
May
20
May
21
May
22
May
25
May
27
May
29
May
28

What is a 'vector embedding' and why do they capture meaning? And how can that be used to build a search system?
Why do we need a vector database? And why are they different from traditional search engines? Should they be different?
Offer a concrete and concise explanation of how you will help students understand and apply this lesson.
Live sessions
2-4 hrs / week
Three hours of courses. Optional guest talks from industry experts every week. Office hours to deep dive into your problems with your instructor
Mon, May 18
5:00 PM—6:30 PM (UTC)
Tue, May 19
4:00 PM—5:00 PM (UTC)
Wed, May 20
5:00 PM—6:30 PM (UTC)
Projects
1-2 hrs / week
Optional labs to expand your knowledge and deliver into your team's codebase
Walking through exactly why / how agents change the game for search+RAG
Hamel Hussain + Doug Turnbull discuss RAG evaluation
$1,450
USD
5 days left to enroll