Quick references for your next loop.
Compact concept cards covering the 14 topics AI engineering interviewers test most. Each sheet fits a focused 10-minute review before a real interview.
LLM Fundamentals
Core concepts behind modern large language models, from transformer architecture to inference. Interviewers test whether you can explain what happens inside the model at a mechanistic level.
Prompt Engineering
Techniques for reliably controlling LLM behavior through prompt design. Interviewers look for systematic thinking about instruction clarity, robustness, and measurable evaluation.
RAG
Patterns for grounding LLM responses in retrieved documents. Interviewers probe your ability to identify and fix quality bottlenecks across the retrieval and generation pipeline.
AI Agents
Design patterns for LLM-driven agents that plan, use tools, and complete multi-step tasks. Interviewers focus on reliability, failure modes, and evaluation of open-ended loops.
AI Agent Reliability
Production patterns for building agents that survive the real world — idempotency, audit trails, approval gates, and clean architecture boundaries. Interviewers test whether you know how agents fail and how to design around those failures.
Fine-Tuning
Techniques for adapting pretrained models to new tasks or domains. Interviewers test your ability to choose the right adaptation strategy and diagnose training problems.
Vector DBs
How to store, index, and retrieve dense vector representations at scale. Interviewers test your understanding of ANN tradeoffs, embedding model selection, and operational concerns.
AI System Design
End-to-end architecture of AI-powered systems at production scale. Interviewers expect you to reason about serving, caching, multi-tenancy, fallbacks, and observability.
LLMOps
Operational practices for shipping and maintaining LLM-powered systems reliably. Interviewers look for experience with rollout safety, cost control, and regression prevention.
Evaluation
Rigorous methods for measuring LLM system quality. Interviewers probe your ability to design reliable evals, choose appropriate metrics, and avoid evaluation pitfalls.
Safety & Ethics
Principles and practices for building AI systems that are safe, fair, and trustworthy. Interviewers test your ability to identify risks and implement mitigations in real products.
Multimodal AI
Systems that process and generate multiple data modalities — images, audio, and text. Interviewers focus on fusion strategies, modality-specific failure modes, and latency.
AI Infrastructure
Hardware, orchestration, and cost strategies for running AI workloads at scale. Interviewers test your ability to design systems that remain performant and cost-efficient under load.
Coding
Engineering fundamentals for building AI systems in production Python. Interviewers often give live coding tasks covering async I/O, streaming, type safety, and API integration.
Behavioral
Soft-skill and leadership questions interviewers use to assess culture fit and engineering maturity. Prepare specific stories using the STAR framework for each focus area.
Put the concepts to work.
Reading cheatsheets is half the job. Run a timed mock interview to practise explaining these concepts under pressure.