AI for Developers
LLMs, prompting, evaluation, responsible AI.
Overview
Apply LLMs responsibly in apps: prompting, evaluation and safety. Build small, reliable AI features.
Learn retrieval, context windows, latency/cost trade‑offs, and when to prefer deterministic alternatives.
Syllabus
- LLM capabilities/limits and responsible AI basics
- Prompt patterns, system prompts and guardrails
- Evaluation: golden sets, metrics and regressions
- Retrieval‑augmented generation and chunking
- Tools, function calling and deterministic fallbacks
- Latency/cost controls, caching and timeouts
- Safety: PII, abuse, jailbreaks and logging
- Hands‑on: add an AI helper with tests