· Toggle for focused reading

2026 Edition
The complete path — no shortcuts

Generative
AI Engineer
Roadmap

From zero to production-grade GenAI engineer in 2026. LLMs, RAG, fine-tuning, agents, MLOps — everything you need with free YouTube resources for every phase.

"Every company is becoming an AI company. The engineer who can ship real GenAI products — not just call APIs — is the most valuable person in the room right now."
— The Boring Education Team
9–15
Months to job-ready
10
Phases to master
45+
Free YT resources
Career ceiling

Start Here — Before You Touch Any Model

1
Weeks 1–3
Phase 01 · Python for AI
Python — The Language of AI, Non-Negotiable
Python is the lingua franca of AI. Master it deeply before touching any model or framework. Cover data types, OOP, list comprehensions, generators, decorators, virtual environments, and async/await. Then learn the scientific stack: NumPy for tensors, Pandas for data wrangling, and Matplotlib for visualization. Build 5 CLI tools from scratch before moving on — no shortcuts here.
Non-negotiable Python 3.12+ NumPy Pandas Matplotlib Async/Await Jupyter Notebooks
2
Weeks 3–6
Phase 02 · Math for AI
Linear Algebra, Calculus & Probability
You don't need a PhD — but you must understand the math behind what models do. Cover vectors, matrices, matrix multiplication, eigenvalues, and dot products (linear algebra). Understand derivatives, gradients, and the chain rule (calculus for backprop). Learn probability: Bayes' theorem, distributions, entropy, and KL divergence. Use 3Blue1Brown — nothing explains this better.
Foundation Linear Algebra Calculus & Gradients Probability Statistics Entropy & KL Divergence
3
Weeks 6–12
Phase 03 · Machine Learning Fundamentals
Classical ML Before Deep Learning — Always
Understand supervised vs unsupervised learning, train/val/test splits, overfitting/underfitting, bias-variance tradeoff. Master: linear regression, logistic regression, decision trees, random forests, gradient boosting (XGBoost), and k-means clustering. Learn scikit-learn deeply. Build real projects: house price predictor, spam classifier, customer churn model. These fundamentals make you a better GenAI engineer — they give you intuition models don't.
Core intuition scikit-learn Supervised Learning Gradient Boosting Cross-validation Feature engineering Model evaluation
🧮
Don't skip the math. Every GenAI engineer who skips math hits a wall at fine-tuning, training stability, or prompt engineering theory. You don't need to derive equations — but understanding why gradients flow the way they do separates engineers who debug models from those who just pray they work.

Neural Networks, PyTorch & The Transformer

4
Weeks 10–18
Phase 04 · Deep Learning with PyTorch
Neural Networks from Scratch — Then PyTorch
Start by building a neural network from scratch in NumPy — forward pass, loss function, backpropagation, gradient descent. Then move to PyTorch: tensors, autograd, nn.Module, DataLoader, optimizers, training loops. Build CNNs for vision and RNNs/LSTMs for sequences. Understand batch normalization, dropout, learning rate schedules, and GPU training. Complete Andrej Karpathy's Neural Networks: Zero to Hero — it is the best deep learning course ever made for engineers.
Core skill PyTorch Backpropagation CNNs RNNs / LSTMs GPU training (CUDA) Training loops Regularization
5
Weeks 16–24
Phase 05 · Transformers & LLM Architecture
Attention Is All You Need — Understand It Deeply
The transformer architecture powers every major LLM. Understand self-attention, multi-head attention, positional encoding, encoder-decoder vs decoder-only architectures, tokenization (BPE, WordPiece), and the scaling laws that govern LLM behavior. Read the original "Attention Is All You Need" paper. Implement a mini GPT from scratch following Karpathy's makemore series. Understand BERT vs GPT paradigms, context windows, KV cache, and how inference works at scale.
Essential Self-Attention Transformers Tokenization / BPE GPT Architecture Scaling Laws KV Cache Context Windows

🤖
Build GPT from scratch — no skipping. Engineers who watch tutorials but never implement from scratch can't debug production models. Karpathy's "Let's build GPT" is 2 hours that will fundamentally change how you think about LLMs. Do it once. Then do it again without watching.
Model Family Best For Access Companies / Use Cases
OpenAI GPT-4o General reasoning, multimodal, APIs API (paid) Startups, enterprise SaaS, copilots
Meta Llama 3.x Open source, fine-tuning, on-prem Free (open weights) Self-hosted apps, regulated industries
Google Gemini Long context, multimodal, Google Cloud API (free tier) Document AI, Google ecosystem apps
Mistral / Mixtral Efficient, fast inference, MoE Open + API Edge AI, cost-sensitive production

LLM APIs, RAG Systems & Prompt Engineering

6
Weeks 20–28
Phase 06 · LLM APIs & Prompt Engineering
Working with LLMs in Production — APIs & Prompts
Learn to integrate LLMs via APIs: OpenAI SDK, Anthropic SDK, Google Gemini SDK, and the unified LiteLLM layer. Master prompt engineering: zero-shot, few-shot, chain-of-thought (CoT), ReAct, structured outputs, and prompt chaining. Understand token limits, pricing, rate limiting, and streaming responses. Use LangChain and LlamaIndex for orchestration. Build a full Q&A bot, a summarizer, and a classification pipeline before moving on.
Production first OpenAI SDK Anthropic SDK LangChain LlamaIndex Chain-of-Thought Structured Output Streaming
7
Weeks 24–34
Phase 07 · RAG Systems
Retrieval-Augmented Generation — The Most In-Demand Skill
RAG is the backbone of 80% of enterprise AI products. Learn vector embeddings: what they are, how to generate them (OpenAI embeddings, sentence-transformers), and cosine similarity. Master vector databases: Pinecone, Weaviate, Chroma, pgvector. Build a full RAG pipeline: chunk documents, embed, store, retrieve, re-rank, and generate. Understand advanced RAG: hybrid search (BM25 + semantic), query rewriting, HyDE, contextual compression, and evaluation with RAGAS. Build a document Q&A system over a PDF corpus as a portfolio project.
Most in-demand Vector Embeddings Pinecone / Chroma pgvector Hybrid Search Re-ranking RAGAS evaluation Document chunking

🔍
RAG quality lives and dies by chunking strategy. Most engineers get RAG working but wonder why answers are bad. The culprit is almost always poor chunking — too large, too small, or no overlap. Learn semantic chunking, recursive splitting, and always evaluate with RAGAS metrics before shipping to production.
Database Best For Free Tier When to Use
Chroma Local dev, prototyping, open source Yes (self-hosted) Learning RAG, small projects, hackathons
Pinecone Managed, production-scale vector search Yes (limited) Production apps, startup MVPs
pgvector SQL + vector search in one DB Yes (open source) Full-stack apps already using Postgres
Weaviate Multimodal, hybrid search, graphs Yes (cloud free) Complex enterprise RAG pipelines

Fine-Tuning, Agents & Production AI

8
Month 6–9 (Ongoing)
Phase 08 · Fine-Tuning & Model Customization
Fine-Tune Open-Source LLMs for Real Use Cases
When prompting isn't enough, you fine-tune. Learn supervised fine-tuning (SFT) with Hugging Face Transformers and TRL. Understand LoRA / QLoRA — parameter-efficient fine-tuning that runs on a single GPU. Use Unsloth for 2× faster training. Master instruction tuning datasets, RLHF basics, DPO (Direct Preference Optimization), and model merging. Work with Llama 3, Mistral, and Gemma. Deploy fine-tuned models with Ollama and vLLM for fast inference.
Senior-level Hugging Face LoRA / QLoRA Unsloth RLHF / DPO Ollama vLLM Model merging PEFT
9
Month 7–11 (Ongoing)
Phase 09 · AI Agents & Multi-Agent Systems
Build Autonomous Agents That Take Real Actions
Agents are the next frontier. Understand the ReAct loop (Reason + Act), tool calling / function calling, and memory (in-context, episodic, semantic). Build agents with LangGraph for complex stateful workflows and CrewAI / AutoGen for multi-agent teams. Learn agent memory patterns, tool use (web search, code execution, APIs), self-reflection loops, and guardrails. Build a research agent, a code generation agent, and a multi-agent pipeline for a real problem. Evaluate agent performance — this is the hardest part.
Cutting edge LangGraph CrewAI AutoGen Function calling Agent memory Tool use Multi-agent Guardrails

🔗 Tool Calling
Structured output where the model returns JSON specifying a function name and args. Your code executes it and feeds results back. Foundation of every production agent.
🧠 Semantic Caching
Cache LLM responses by semantic similarity, not exact match. Reduces API costs 40–70% for production apps. Use GPTCache or Redis with vector similarity.
📊 LLM Evaluation
You cannot improve what you don't measure. Use LLM-as-judge, RAGAS for RAG pipelines, and frameworks like DeepEval and PromptFoo to systematically evaluate outputs.
🔒 AI Guardrails
Production AI needs safety layers. Use Guardrails AI, NVIDIA NeMo Guardrails, or custom output validators to prevent hallucinations, PII leaks, and harmful outputs reaching users.

MLOps, Getting Hired & Building in Public

10
Month 9–15 (Interview Prep)
Phase 10 · MLOps, Deployment & Getting Hired
Ship AI Products + Land the Job
Learn MLOps: experiment tracking with MLflow / Weights & Biases, model versioning, model registries, and CI/CD for ML pipelines. Deploy models as REST APIs using FastAPI + Docker. Learn inference optimization: quantization (GGUF, GPTQ), model distillation, and batching strategies. Use Hugging Face Spaces and Streamlit for demos. Build 2–3 complete AI products with live demos: a RAG chatbot, an AI agent, and a fine-tuned model endpoint. Document everything. Your GitHub and your demo links close offers — not your resume alone.
Job-ready MLflow / W&B FastAPI + Docker Streamlit / Gradio Quantization HF Spaces CI/CD for ML GitHub portfolio

What to Learn & When — Full Timeline
🟥 Month 1–3
Python (NumPy, Pandas)
Linear Algebra & Calculus
Classical ML (scikit-learn)
Jupyter Notebooks
Feature Engineering basics
Git & GitHub workflows
Stats & Probability
🟧 Month 4–7
PyTorch & Deep Learning
Transformer architecture
LLM APIs (OpenAI, Gemini)
Prompt Engineering
LangChain / LlamaIndex
RAG pipelines
Vector databases
🟩 Month 8–15
Fine-tuning (LoRA/QLoRA)
AI Agents (LangGraph)
MLOps (MLflow, W&B)
Model inference (vLLM)
FastAPI + Docker deploys
LLM evaluation (RAGAS)
Build in public

The Boring GenAI Routine That Actually Works
1 Karpathy / 3Blue1Brown video watched with notebook open — implement what you see
30 min on Hugging Face docs or one paper summary (Papers With Code)
Push at least 1 commit to your AI project on GitHub — no matter how small
Write down 1 thing you don't understand — go deep on it tomorrow
Share 1 thing you built or learned on LinkedIn or Twitter — build in public

Best Free YouTube Channels for GenAI

📺 Andrej Karpathy
The single best resource on the internet for understanding LLMs from first principles. Built GPT, was OpenAI's founding member, Tesla AI director. Every video is gold.
📺 3Blue1Brown
Grant Sanderson's visual explanations of neural networks, transformers, and attention mechanisms are unmatched. If you're struggling with math or architecture intuition, start here.
📺 StatQuest with Josh Starmer
The clearest explanations of ML concepts, statistics, and deep learning on YouTube. Every StatQuest video is structured to build intuition before equations — perfect for self-learners.
📺 Sam Witteveen
The most practical GenAI engineering channel — LangChain, agents, RAG, fine-tuning tutorials that are always up to date with the latest models and frameworks. Build-first approach.
📺 Yannic Kilcher
Deep paper reading sessions on the latest AI research — GPT-4, diffusion models, alignment techniques, and more. If you want to understand what the frontier looks like, follow Yannic.
📺 AI Jason / Matt Williams
Hands-on tutorials for Ollama, local LLMs, agents, and practical GenAI apps. Perfect for engineers who want to run models locally and build real production projects fast.

DSA Yatra — Daily practice Prep Yatra — Interview tracker Tech Yatra — Learning roadmaps Resume Yatra — ATS-ready resume Shiksha — Free courses YouFocus — Distraction-free YT Interview Prep — Question banks Community — Peer learning
Your GenAI Journey Starts Now 🤖
The best time to start was a year ago. The second best time is today.
Consistency over 12 months beats any bootcamp or degree. Go build.
→ theboringeducation.com