ML Engineer
Interview-ready on fine-tuning, alignment, and LLM evaluation
ML engineer interviews now include deep LLM questions: "When would you fine-tune vs. use RAG?", "Explain LoRA," "How do you evaluate a fine-tuned model?" This path covers transformer math, PEFT techniques, training data curation, alignment methods (RLHF, DPO), and evaluation frameworks — everything you need for ML engineering interviews in the LLM era.
Your Learning Path
A step-by-step roadmap from foundations to mastery. Follow this sequence for the most effective learning experience.
Modules
81 free module to get you started, plus 7 premium deep-dives.
ML Engineer Roadmap
The complete learning path for ML engineers working with LLMs: from transformer fundamentals to production fine-tuning and alignment. Understand the difference between GenAI engineering and ML engineering roles.
LLM Architecture Deep Dive
Transformer architecture at the mathematical level: self-attention equations, multi-head attention, positional encodings (RoPE, ALiBi), layer normalization, feed-forward networks, and how modern LLMs (GPT, Llama, Claude) differ architecturally.
The Fine-Tuning Decision
When to fine-tune vs. use prompting vs. RAG. Cost-benefit analysis frameworks, data requirements estimation, compute budgeting, and a decision tree for choosing the right approach for your use case.
Parameter-Efficient Fine-Tuning (PEFT)
Deep dive into LoRA, QLoRA, prefix tuning, adapters, and IA3. Understand the math behind low-rank adaptation, how to choose rank and alpha hyperparameters, and when each PEFT method shines.
Training Data for Fine-Tuning
Building high-quality fine-tuning datasets: data collection strategies, annotation guidelines, quality filtering, synthetic data generation, data formatting (Alpaca, ShareGPT, chat templates), and dataset evaluation.
Running Fine-Tuning Jobs
Hands-on fine-tuning execution: HuggingFace Transformers + TRL, Axolotl, cloud GPU provisioning (Lambda Labs, RunPod, AWS), hyperparameter tuning, distributed training basics, and experiment tracking with W&B.
Alignment: RLHF, DPO, ORPO
How models learn to follow instructions and be helpful. Reward model training, PPO for RLHF, Direct Preference Optimization (DPO), Odds Ratio Preference Optimization (ORPO), constitutional AI, and building preference datasets.
Model Evaluation
Comprehensive LLM evaluation: automated benchmarks (MMLU, HumanEval, MT-Bench), human evaluation protocols, task-specific metrics, LLM-as-judge, regression testing, and building evaluation pipelines for fine-tuned models.
Start Free — No Account Required
These foundational resources are free for everyone. Build your AI literacy before diving into persona-specific modules.
Unlock All 7 Premium Modules
Get full access to every ML Engineer module — plus all other GenAI personas, DSA content, and System Design content with a single subscription.
View Pricing