LLM Engineering - Foundations to SLMs

4.6ย (13)

ยท

5 Weeks

ยท

Cohort-based Course

Master language models and embedding models through training, fine-tuning, aligning, distilling, and merging transformer architectures!

This course is popular

4 people enrolled last week.

Course overview

Understand LLMs from first principles; build custom Small Language Models (SLMs)

๐Ÿง‘โ€๐Ÿ’ป Language Model Engineering refers to the evolving set of best practices for training, fine-tuning, and aligning LLMs to optimize their use and function to balance performance with efficiency.


These best practices have formed the basis for the LLMs, or Large Language Models (a.k.a. Foundation Models) and Small Language Models (SLMs) of today.


๐Ÿค– Whether youโ€™re looking at OpenAIโ€™s GPT series, Anthropic Claude, Grok, Google Gemini, Mistral, or any other model provider, the core underlying transformer architectures are similar, as are the training and tuning methods.


When chasing after high scores on LLM benchmarks or creating state-of-the-art SLMs, techniques like Model Merging and Distillation are important as well.


๐Ÿซ This course will provide you with the foundational concepts and code to train, fine-tune, and align state-of-the-art SLMs and LLMs using industry-standard and emerging approaches from the open-source edge heading into 2025.


๐Ÿค“ Become the expert in your organization on all things training (pretraining, post-training), fine-tuning (supervised fine-tuning, instruction tuning, chat tuning, etc.), alignment (PPO, DPO, etc.), Small Language Models (SLMs), Model Merging, Distillation, and more!

Who is this course for

01

Data scientists, researchers, or engineers experienced with classic Machine Learning who want to build custom Small Language Models (SLMs).

02

Product managers who need to understand the core concepts and code behind custom SLMs to effectively lead their product teams.

03

Stakeholders who aim to train, fine-tune, or align proprietary SLMs for their customers in 2025.

By the end of this course, you will be able to:

๐Ÿฆพ Build, train, and evaluate Language Models using transformer variants (GPT, BERT, BART)


๐Ÿง Calculate self-attention and understand the latest implementations (Flash Attention, FA2)

๐Ÿ”  Demystify embedding layers, embedding representations, pre-trained vs. learned embeddings, ROPE, and more!

๐Ÿช™ Decode embedding space representations for optimal next-token prediction

๐Ÿ”กย Build, train, and evaluate embedding models (like those in ๐Ÿค— Sentence Transformers)

๐Ÿš‡ย Complete unsupervised and continued pretraining of LLMs and SLMs from scratch

๐Ÿš‰ย Fine-tune pre-trained LMs for instruction-following, chat, and more via parameter-efficient methods

๐Ÿ›ค๏ธ Align LMs to balance helpfulness with harmlessness and other criteria (RLXF, DPO)

๐Ÿš€ Explore frontiers of Language Models (Mixture-of- approaches, Model Merging, alternative fine-tuning, and more!)

This course includes

Interactive live sessions

Lifetime access to course materials

78 in-depth lessons

Direct access to instructor

8 projects to apply learnings

Guided feedback & reflection

Private community of peers

Course certificate upon completion

Maven Satisfaction Guarantee

This course is backed by Mavenโ€™s guarantee. You can receive a full refund within 14 days after the course ends, provided you meet the completion criteria in our refund policy.

Course syllabus

Week 1

Feb 11โ€”Feb 16

    ๐Ÿˆ Cohort Kick-Off!

    6 items

    ๐Ÿค– The Transformer Paper: Attention is All You Need

    8 items

Week 2

Feb 17โ€”Feb 23

    ๐Ÿง Attention

    6 items

    ๐Ÿ”  Embeddings

    10 items

Week 3

Feb 24โ€”Mar 2

    ๐Ÿช™ Next-Token Prediction

    9 items

    ๐Ÿ”ก Embedding Models

    9 items

Week 4

Mar 3โ€”Mar 9

    ๐Ÿš‡ Pretraining

    10 items

    ๐Ÿš‰ Fine Tuning

    11 items

Week 5

Mar 10โ€”Mar 13

    ๐Ÿ›ค๏ธ Alignment

    10 items

    ๐ŸŒŒ LLM Engineering Frontiers

    7 items

4.6ย (13 ratings)

What students are saying

LLM Engineering vs. AI Engineering: What's the Difference?

Prerequisites

A background in fundamental Machine Learning and Deep Learning

Understanding supervised learning, unsupervised learning, and neural network architectures is required. Introductory NLP and Computer Vision knowledge is encouraged. Not sure where to start? Read this.

A ability to program in Python within a Jupyter Notebook Environment

Understand basic Python syntax and constructs. You should be comfortable training and evaluating simple ML & DL models using test, train, and dev sets. Not sure where to start? Look here.

Free resource

The LLM Engineering Onramp

Youโ€™ll find brief introductions to core concepts weโ€™ll cover in more depth in the class - this is meant to get the learning juices flowing with bite-sized training materials from Dr. Greg and The Wiz.

Start learning for free now!

Free resource

๐Ÿ“… Detailed Schedule!

Understand how everything comes together in the course to provide a holistic overview of the how LLMs are engineered.


Get all the details about the assignments, associated papers, and key concepts you'll learn!

Send me the deets โœŒ๏ธ

Meet your instructors

"Dr. Greg" Loughnane

"Dr. Greg" Loughnane

Co-Founder & CEO @ AI Makerspace

In 2023, "The Wiz ๐Ÿช„" a.k.a. "The LLM Wizard ๐Ÿช„" and I created theย LLM Engineering: The Foundationsย andย LLM Ops: LLMs in Productionย courses. In 2024, we launched The AI Engineering Bootcamp.


From 2021-2023 I led the product & curriculum team at FourthBrain (Backed by Andrew Ng's AI Fund) to build industry-leading online bootcamps in ML Engineering and ML Operations (MLOps).


Previously, I worked as an AI product manager, university professor, data science consultant, AI startup advisor, and ML researcher; TEDx & keynote speaker, lecturing since 2013.


๐Ÿ‘จโ€๐Ÿซ Learn with us free onย YouTube!

๐Ÿ‘จโ€๐Ÿ’ผ Connect with me onย LinkedIn!

Chris "The Wiz ๐Ÿช„" Alexiuk

Chris "The Wiz ๐Ÿช„" Alexiuk

Co-Founder & CTO @ AI Makerspace

During the day, I work as a Developer Advocate for NVIDIA. Previously, I worked with Greg at FourthBrain (Backed by Andrew Ng's AI Fund) on MLE and MLOps courses, and on a few Deeplearning.ai events!


A former founding MLE and data scientist, these days you can find me cranking out Machine Learning and LLM content!


My motto is "Build, build, build!", and I'm excited to get building with all of you!


๐Ÿ‘จโ€๐Ÿซ YouTube:ย AI Makerspace Official,ย My Personal

๐Ÿ‘จโ€๐Ÿ’ผ Connect with me onย LinkedIn!

A pattern of wavy dots

Join an upcoming cohort

LLM Engineering - Foundations to SLMs

Cohort 4

$1,999

Dates

Feb 11โ€”Mar 14, 2025

Payment Deadline

Feb 10, 2025

Course schedule

4-6 hours per week

  • Class!

    Tuesdays & Thursdays, 4:00-6:00pm PT

    • Feb 11th, 13th
    • Feb 18th, 20th
    • Feb 25th, 27th
    • Mar 4th, 6th
    • Mar 11th, 13th
    • Mar 18th, 20th
  • Weekly Programming Projects

    2-4 hours per week

    Each class period, we will get hands-on with Python coding homework!

  • Office Hours

    Tuesdays and Fridays

    Dr. Greg: Thursdays @ 8 AM PT

    The Wiz ๐Ÿช„: Fridays @ 3 PM PT

  • Peer Support Sessions

    Varies

    Peers supporters are here to serve as your destination group; they're available for homework help, concept deep dives, and to hang out with!


    Find them live in class, on Discord, and even in their own dedicated weekly sessions!

Build ๐Ÿ—๏ธ, ship ๐Ÿšข, and share ๐Ÿš€ like a legend.

Build ๐Ÿ—๏ธ, ship ๐Ÿšข, and share ๐Ÿš€ like a legend.

Hands-on learning at the LLM edge

We teach concepts AND code. Never one or the other. AI has accelerated so quickly that anyone who's been a manager the last few years has not yet seen the code for themselves.

Pair programming made fun and easy

Grow your network. Build, ship, and share with a community! Build relationships with peers and instructional staff to unlock opportunities in the years ahead.

Find fellow travelers for your journey

Join a community of like-minded AI practitioners who are all in on Generative AI, and who are heading down the same career path as you are.

Frequently Asked Questions

A pattern of wavy dots

Join an upcoming cohort

LLM Engineering - Foundations to SLMs

Cohort 4

$1,999

Dates

Feb 11โ€”Mar 14, 2025

Payment Deadline

Feb 10, 2025

$1,999

4.6ย (13)

ยท

5 Weeks