AI
AI Engineering
Build with large language models: prompts, RAG, function calling, evaluation, deploy.
18 skills
6 months
Advanced
You'll cover practical prompt engineering, embeddings and retrieval-augmented generation (RAG), tool use / function calling, multi-step agents, response evaluation, and the safety patterns that keep the bill and data leaks under control.
Outcomes: AI engineer, AI product engineer, consultant for AI integrations.
Foundation
Core
3
Designing System Prompts
Persona, scope, refusal patterns, output format. Building the model's personality.
25 minutes
4
Few-Shot Prompting
Showing examples instead of describing — when 3 demos beat 30 lines of instructions.
20 minutes
5
Structured Outputs (JSON)
Forcing valid JSON with schemas. Tool/function calling, JSON mode, validation.
25 minutes
7
The Claude API
Anthropic SDK, messages API, system, max_tokens, stop_sequences. The essentials.
30 minutes
8
The OpenAI API
Chat Completions API, Responses API, tools, comparison points with Claude.
25 minutes
10
Streaming Responses
SSE, partial-text rendering, perceived speed wins. Implementing it cleanly.
20 minutes
11
Embeddings Fundamentals
Turning text into vectors, cosine similarity, picking models, dimensionality.
25 minutes
12
Vector Databases
pgvector, Qdrant, Pinecone. ANN indexes, metadata filtering, hybrid search.
30 minutes
17
Cost Optimization
Smaller models, prompt caching, tighter outputs, batch APIs. Where the bill goes.
20 minutes
Advanced
6
Tool Use / Function Calling
Letting the model call your code: tool schemas, multi-turn flow, error handling.
35 minutes
9
Prompt Caching
Reusing the static part of your prompts to cut latency and cost dramatically.
20 minutes
13
RAG Basics
Chunk, embed, retrieve, ground. The architecture that makes LLMs answer your data.
35 minutes
14
Advanced RAG Retrieval
Reranking, query rewriting, hybrid retrieval, when to add a small tuned reranker.
30 minutes
15
Multi-Step Agents
Loops, planning, reflection, tool selection. Pitfalls of "let the model figure it out".
35 minutes
16
Evaluating LLM Outputs
Datasets, golden answers, LLM-as-judge, regression suites. Treating prompts like code.
30 minutes
18
Safety and Guardrails
Input filtering, output checks, jailbreak resistance, abuse detection.
25 minutes