The complete path — no shortcuts
Generative
AI Engineer
Roadmap
From zero to production-grade GenAI engineer in 2026. LLMs, RAG, fine-tuning, agents, MLOps — everything you need with free YouTube resources for every phase.
"Every company is becoming an AI company. The engineer who can ship real GenAI products — not
just call APIs — is the most valuable person in the room right now."
— The Boring Education Team
9–15
Months to job-ready
10
Phases to master
45+
Free YT resources
∞
Career ceiling
theboringeducation.com · Free Tech Education for
Everyone
01
Foundation Layer
Start Here — Before You Touch Any Model
1
Weeks 1–3
Phase 01 · Python for AI
Python — The Language of AI, Non-Negotiable
Python is the lingua franca of AI. Master it deeply before touching
any model or framework. Cover data types, OOP, list comprehensions, generators,
decorators, virtual environments, and async/await. Then learn the scientific stack:
NumPy for tensors, Pandas for data wrangling, and Matplotlib for visualization. Build
5 CLI tools from scratch before moving on — no shortcuts here.
Non-negotiable
Python 3.12+
NumPy
Pandas
Matplotlib
Async/Await
Jupyter Notebooks
2
Weeks 3–6
Phase 02 · Math for AI
Linear Algebra, Calculus & Probability
You don't need a PhD — but you must understand the math behind what
models do. Cover vectors, matrices, matrix multiplication, eigenvalues, and dot
products (linear algebra). Understand derivatives, gradients, and the chain rule
(calculus for backprop). Learn probability: Bayes' theorem, distributions, entropy,
and KL divergence. Use 3Blue1Brown — nothing explains this better.
Foundation
Linear Algebra
Calculus & Gradients
Probability
Statistics
Entropy & KL Divergence
3
Weeks 6–12
Phase 03 · Machine Learning Fundamentals
Classical ML Before Deep Learning — Always
Understand supervised vs unsupervised learning, train/val/test
splits, overfitting/underfitting, bias-variance tradeoff. Master: linear regression,
logistic regression, decision trees, random forests, gradient boosting (XGBoost), and
k-means clustering. Learn scikit-learn deeply. Build real projects: house price
predictor, spam classifier, customer churn model. These fundamentals make you a better
GenAI engineer — they give you intuition models don't.
Core intuition
scikit-learn
Supervised Learning
Gradient Boosting
Cross-validation
Feature engineering
Model evaluation
🧮
Don't skip the math. Every GenAI engineer who skips math
hits a wall at fine-tuning, training stability, or prompt engineering theory. You don't need
to derive equations — but understanding why gradients flow the way they do separates
engineers who debug models from those who just pray they work.
theboringeducation.com
02 / 07
Deep Learning & Transformers
Neural Networks, PyTorch & The Transformer
4
Weeks 10–18
Phase 04 · Deep Learning with PyTorch
Neural Networks from Scratch — Then PyTorch
Start by building a neural network from scratch in NumPy — forward
pass, loss function, backpropagation, gradient descent. Then move to
PyTorch: tensors, autograd, nn.Module, DataLoader, optimizers,
training loops. Build CNNs for vision and RNNs/LSTMs for sequences. Understand batch
normalization, dropout, learning rate schedules, and GPU training. Complete Andrej
Karpathy's Neural Networks: Zero to Hero — it is the best deep learning course ever
made for engineers.
Core skill
PyTorch
Backpropagation
CNNs
RNNs / LSTMs
GPU training (CUDA)
Training loops
Regularization
5
Weeks 16–24
Phase 05 · Transformers & LLM Architecture
Attention Is All You Need — Understand It Deeply
The transformer architecture powers every major LLM. Understand
self-attention, multi-head attention, positional encoding, encoder-decoder vs
decoder-only architectures, tokenization (BPE, WordPiece), and the scaling laws that
govern LLM behavior. Read the original "Attention Is All You Need" paper. Implement a
mini GPT from scratch following Karpathy's makemore series. Understand BERT vs GPT
paradigms, context windows, KV cache, and how inference works at scale.
Essential
Self-Attention
Transformers
Tokenization / BPE
GPT Architecture
Scaling Laws
KV Cache
Context Windows
🤖
Build GPT from scratch — no skipping. Engineers who watch
tutorials but never implement from scratch can't debug production models. Karpathy's "Let's
build GPT" is 2 hours that will fundamentally change how you think about LLMs. Do it once. Then
do it again without watching.
LLM Model Families — Know Your Landscape
| Model Family | Best For | Access | Companies / Use Cases |
|---|---|---|---|
| OpenAI GPT-4o | General reasoning, multimodal, APIs | API (paid) | Startups, enterprise SaaS, copilots |
| Meta Llama 3.x | Open source, fine-tuning, on-prem | Free (open weights) | Self-hosted apps, regulated industries |
| Google Gemini | Long context, multimodal, Google Cloud | API (free tier) | Document AI, Google ecosystem apps |
| Mistral / Mixtral | Efficient, fast inference, MoE | Open + API | Edge AI, cost-sensitive production |
theboringeducation.com
03 / 07
Applied GenAI Engineering
LLM APIs, RAG Systems & Prompt Engineering
6
Weeks 20–28
Phase 06 · LLM APIs & Prompt Engineering
Working with LLMs in Production — APIs & Prompts
Learn to integrate LLMs via APIs: OpenAI SDK, Anthropic SDK,
Google Gemini SDK, and the unified LiteLLM layer. Master prompt
engineering: zero-shot, few-shot, chain-of-thought (CoT), ReAct, structured outputs,
and prompt chaining. Understand token limits, pricing, rate limiting, and streaming
responses. Use LangChain and LlamaIndex for
orchestration. Build a full Q&A bot, a summarizer, and a classification pipeline
before moving on.
Production first
OpenAI SDK
Anthropic SDK
LangChain
LlamaIndex
Chain-of-Thought
Structured Output
Streaming
7
Weeks 24–34
Phase 07 · RAG Systems
Retrieval-Augmented Generation — The Most In-Demand Skill
RAG is the backbone of 80% of enterprise AI products. Learn
vector embeddings: what they are, how to generate them (OpenAI
embeddings, sentence-transformers), and cosine similarity. Master
vector databases: Pinecone, Weaviate, Chroma, pgvector. Build a
full RAG pipeline: chunk documents, embed, store, retrieve, re-rank, and generate.
Understand advanced RAG: hybrid search (BM25 + semantic), query rewriting, HyDE,
contextual compression, and evaluation with RAGAS. Build a document Q&A system
over a PDF corpus as a portfolio project.
Most in-demand
Vector Embeddings
Pinecone / Chroma
pgvector
Hybrid Search
Re-ranking
RAGAS evaluation
Document chunking
🔍
RAG quality lives and dies by chunking strategy. Most
engineers get RAG working but wonder why answers are bad. The culprit is almost always poor
chunking — too large, too small, or no overlap. Learn semantic chunking, recursive splitting,
and always evaluate with RAGAS metrics before shipping to production.
Vector Database Comparison
| Database | Best For | Free Tier | When to Use |
|---|---|---|---|
| Chroma | Local dev, prototyping, open source | Yes (self-hosted) | Learning RAG, small projects, hackathons |
| Pinecone | Managed, production-scale vector search | Yes (limited) | Production apps, startup MVPs |
| pgvector | SQL + vector search in one DB | Yes (open source) | Full-stack apps already using Postgres |
| Weaviate | Multimodal, hybrid search, graphs | Yes (cloud free) | Complex enterprise RAG pipelines |
theboringeducation.com
04 / 07
Advanced GenAI Engineering
Fine-Tuning, Agents & Production AI
8
Month 6–9 (Ongoing)
Phase 08 · Fine-Tuning & Model Customization
Fine-Tune Open-Source LLMs for Real Use Cases
When prompting isn't enough, you fine-tune. Learn supervised
fine-tuning (SFT) with Hugging Face Transformers and
TRL. Understand LoRA / QLoRA — parameter-efficient
fine-tuning that runs on a single GPU. Use Unsloth for 2× faster
training. Master instruction tuning datasets, RLHF basics, DPO (Direct Preference
Optimization), and model merging. Work with Llama 3, Mistral, and Gemma. Deploy
fine-tuned models with Ollama and vLLM for fast
inference.
Senior-level
Hugging Face
LoRA / QLoRA
Unsloth
RLHF / DPO
Ollama
vLLM
Model merging
PEFT
9
Month 7–11 (Ongoing)
Phase 09 · AI Agents & Multi-Agent Systems
Build Autonomous Agents That Take Real Actions
Agents are the next frontier. Understand the ReAct loop
(Reason + Act), tool calling / function calling, and memory (in-context, episodic,
semantic). Build agents with LangGraph for complex stateful workflows
and CrewAI / AutoGen for multi-agent teams. Learn
agent memory patterns, tool use (web search, code execution, APIs), self-reflection
loops, and guardrails. Build a research agent, a code generation agent, and a
multi-agent pipeline for a real problem. Evaluate agent performance — this is the
hardest part.
Cutting edge
LangGraph
CrewAI
AutoGen
Function calling
Agent memory
Tool use
Multi-agent
Guardrails
Essential AI Engineering Patterns
🔗 Tool Calling
Structured output where the model returns JSON specifying a
function name and args. Your code executes it and feeds results back. Foundation of every
production agent.
🧠 Semantic Caching
Cache LLM responses by semantic similarity, not exact match.
Reduces API costs 40–70% for production apps. Use GPTCache or Redis with vector similarity.
📊 LLM Evaluation
You cannot improve what you don't measure. Use LLM-as-judge,
RAGAS for RAG pipelines, and frameworks like DeepEval and PromptFoo to systematically
evaluate outputs.
🔒 AI Guardrails
Production AI needs safety layers. Use Guardrails AI, NVIDIA
NeMo Guardrails, or custom output validators to prevent hallucinations, PII leaks, and
harmful outputs reaching users.
theboringeducation.com
05 / 07
Endgame
MLOps, Getting Hired & Building in Public
10
Month 9–15 (Interview Prep)
Phase 10 · MLOps, Deployment & Getting Hired
Ship AI Products + Land the Job
Learn MLOps: experiment tracking with MLflow /
Weights & Biases, model versioning, model registries, and CI/CD for ML pipelines.
Deploy models as REST APIs using FastAPI + Docker. Learn inference optimization:
quantization (GGUF, GPTQ), model distillation, and batching strategies. Use
Hugging Face Spaces and Streamlit for demos.
Build 2–3 complete AI products with live demos: a RAG chatbot, an AI agent, and a
fine-tuned model endpoint. Document everything. Your GitHub and your demo links close
offers — not your resume alone.
Job-ready
MLflow / W&B
FastAPI + Docker
Streamlit / Gradio
Quantization
HF Spaces
CI/CD for ML
GitHub portfolio
Skill Map
What to Learn & When — Full Timeline
🟥 Month 1–3
Python (NumPy, Pandas)
Linear Algebra & Calculus
Classical ML (scikit-learn)
Jupyter Notebooks
Feature Engineering basics
Git & GitHub workflows
Stats & Probability
🟧 Month 4–7
PyTorch & Deep Learning
Transformer architecture
LLM APIs (OpenAI, Gemini)
Prompt Engineering
LangChain / LlamaIndex
RAG pipelines
Vector databases
🟩 Month 8–15
Fine-tuning (LoRA/QLoRA)
AI Agents (LangGraph)
MLOps (MLflow, W&B)
Model inference (vLLM)
FastAPI + Docker deploys
LLM evaluation (RAGAS)
Build in public
Daily Routine
The Boring GenAI
Routine That Actually Works
1 Karpathy / 3Blue1Brown video watched with notebook open — implement what you see
30 min on Hugging Face docs or one paper summary (Papers With Code)
Push at least 1 commit to your AI project on GitHub — no matter how small
Write down 1 thing you don't understand — go deep on it tomorrow
Share 1 thing you built or learned on LinkedIn or Twitter — build in public
theboringeducation.com
06 / 07
Master Resource List
Best Free YouTube Channels for GenAI
📺 Andrej Karpathy
The single best resource on the internet for understanding LLMs
from first principles. Built GPT, was OpenAI's founding member, Tesla AI director. Every
video is gold.
📺 3Blue1Brown
Grant Sanderson's visual explanations of neural networks,
transformers, and attention mechanisms are unmatched. If you're struggling with math or
architecture intuition, start here.
📺 StatQuest with Josh Starmer
The clearest explanations of ML concepts, statistics, and
deep learning on YouTube. Every StatQuest video is structured to build intuition before
equations — perfect for self-learners.
📺 Sam Witteveen
The most practical GenAI engineering channel — LangChain,
agents, RAG, fine-tuning tutorials that are always up to date with the latest models and
frameworks. Build-first approach.
📺 Yannic Kilcher
Deep paper reading sessions on the latest AI research —
GPT-4, diffusion models, alignment techniques, and more. If you want to understand what
the frontier looks like, follow Yannic.
📺 AI Jason / Matt Williams
Hands-on tutorials for Ollama, local LLMs, agents, and
practical GenAI apps. Perfect for engineers who want to run models locally and build
real production projects fast.
Tools by TBE — Use These
DSA Yatra — Daily practice
Prep Yatra — Interview tracker
Tech Yatra — Learning roadmaps
Resume Yatra — ATS-ready resume
Shiksha — Free courses
YouFocus — Distraction-free YT
Interview Prep — Question banks
Community — Peer learning
Your GenAI Journey Starts Now 🤖
The best time to start was a year ago. The second best time is today.
Consistency over 12 months beats any bootcamp or degree. Go build.
→ theboringeducation.com
Consistency over 12 months beats any bootcamp or degree. Go build.
Find Us Everywhere
© 2026 The Boring Education · Free Tech Education for Everyone
07 / 07