Course overview

How to Design Artificial Intelligence

41 modules
176 lessons
—
Part 1

Course Setup and the Incremental Ladder

  1. Course Setup and the Incremental LadderSign in

  2. Why "Transformers to Intelligence"Sign in

  3. How to Use This CourseSign in

  4. The Incremental Ladder (Step 0 to Step 7)Sign in

  5. The Course LensesSign in

  6. Diagram Legend and Notation TypesSign in

Part 2

Mental Models: Functions, Learning, and Intelligence

  1. Mental Models: Functions, Learning, and IntelligenceSign in

  2. Learning as Function ApproximationSign in

  3. Intelligence as Compression, Prediction, and ControlSign in

  4. Local Optimization, Global BehaviorSign in

Part 3

Mathematical Foundations

  1. Mathematical FoundationsSign in

  2. Linear Algebra for AI: vectors, matrices, tensors, and why geometry is the hidden structure of modelsSign in

  3. Probability and Uncertainty: distributions, expectations, Bayes rule, and calibration as an operational concernSign in

  4. Optimization in Practice: gradients, convex vs non-convex intuition, SGD and common variantsSign in

Part 4

Physics-Inspired Views of Learning

  1. Physics-Inspired Views of LearningSign in

  2. Information Theory as a Lens: entropy, mutual information, and compression-driven interpretations of learningSign in

  3. Energy Landscapes and Statistical Mechanics Intuition: why minima, basins, and temperature metaphors explain training behaviorSign in

  4. Scaling Limits: compute, energy, bandwidth, and the real costs that constrain "bigger models."Sign in

Part 5

Diagramming AI Systems

  1. Diagramming AI SystemsSign in

  2. Computation Graphs and Layers: forward/backward passes as the "wiring diagram" of learning.Sign in

  3. Training vs Inference Paths: data pipelines, feedback loops, and the boundary where behavior is measured.Sign in

  4. System Topologies for AI: model, retrieval, tools, data stores, and users as a coupled system.Sign in

Part 6

Step 0 Representations: Vectors, Embeddings, Tokens

  1. Step 0 Representations: Vectors, Embeddings, TokensSign in

  2. Feature Spaces and Embeddings: one-hot vs dense representations and what "meaning in geometry" implies.Sign in

  3. Tokenization Across Modalities: text, images, and structured inputs as discrete interfaces to continuous models.Sign in

  4. Normalization and Scaling: preprocessing as a stability tool, not a cosmetic step.Sign in

Part 7

Step 0 Models: Linear and Logistic Regression

  1. Step 0 Models: Linear and Logistic RegressionSign in

  2. Linear Regression and MSE: fitting as projection and what linearity limits you to.Sign in

  3. Logistic Regression and Softmax: decision boundaries, cross-entropy, and multi-class classification.Sign in

  4. Training Loops and Optimization Basics: learning rates, SGD, and the earliest forms of regularization.Sign in

Part 8

Step 0 Delivery: Packaging, Serving, and Evaluating Linear Models

  1. Step 0 Delivery: Packaging, Serving, and Evaluating Linear ModelsSign in

  2. From Notebook to API: packaging artifacts, input contracts, and reproducible inference.Sign in

  3. Offline vs Online Evaluation: calibration, drift, and stability as production realities.Sign in

  4. When Linear is Enough: diagnosing "ceiling effects" and knowing when representation learning is required.Sign in

Part 9

Step 1 MLPs and Depth

  1. Step 1 MLPs and DepthSign in

  2. Why Depth Matters: universal approximation versus practical learnability.Sign in

  3. Activations and Initialization: vanishing/exploding gradients as architectural constraints.Sign in

  4. Designing MLPs for Structured Data: inductive bias for tabular and mixed-feature inputs.Sign in

Part 10

Step 1 Convolutions and Local Structure

  1. Step 1 Convolutions and Local StructureSign in

  2. Convolution as an Inductive Bias: receptive fields, weight sharing, and locality.Sign in

  3. Pooling, Stride, and Invariance: what invariance buys and what it breaks.Sign in

  4. Beyond Images: extending convolutional structure to audio, time series, and other domains.Sign in

Part 11

Step 1 Representation Learning and Regularization

  1. Step 1 Representation Learning and RegularizationSign in

  2. Autoencoders and Bottlenecks: learning feature spaces by compressionSign in

  3. Regularization Toolbox: dropout, batch normalization, weight decay, and how they change training dynamicsSign in

  4. Transfer Learning: pretrain then adapt, and why data boundary choices matterSign in

Part 12

Step 2 Recurrent Networks and Temporal Dependencies

  1. Step 2 Recurrent Networks and Temporal DependenciesSign in

  2. RNNs, LSTMs, GRUs: representing state over time and what "memory" means in a modelSign in

  3. Training instability in time: truncated BPTT, gradient pathologies, and practical mitigationSign in

  4. Seq2Seq abstractions: encoder-decoder thinking and where attention first enters as a remedySign in

Part 13

Step 2 Attention Mechanisms

  1. Step 2 Attention MechanismsSign in

  2. Query-Key-Value: attention as differentiable selection and routingSign in

  3. Soft vs hard attention: trade-offs in differentiability, interpretability, and optimizationSign in

  4. Longer contexts: scaling attention and the costs that push you toward transformer-style designsSign in

Part 14

Step 2 Deployment: Architecting and Serving Sequence Models

  1. Step 2 Deployment: Architecting and Serving Sequence ModelsSign in

  2. Streaming vs Offline Inference: boundary choices for latency and correctnessSign in

  3. Memory footprint and sequence length: performance ceilings and practical constraintsSign in

  4. Evaluating sequence tasks: perplexity, BLEU/ROUGE-style metrics, and domain-specific success definitionsSign in

Part 15

Anatomy of a Transformer

  1. Anatomy of a TransformerSign in

  2. Multi-Head Self-Attention: context mixing as the core computational primitiveSign in

  3. Positional encoding: absolute vs relative schemes and what they enable or preventSign in

  4. Feed-forward, residuals, normalization: stability mechanisms and depth scaling behaviorSign in

Part 16

Transformer Variants

  1. Transformer VariantsSign in

  2. Encoder-only, decoder-only, encoder-decoder: matching architecture to task constraintsSign in

  3. Long-context strategies: sparse and linearized attention and the costs they shift elsewhereSign in

  4. Mixture-of-experts and scaling variants: capacity, routing, and operational complexitySign in

Part 17

Training Transformers

  1. Training TransformersSign in

  2. Data and tokenization pipelines: corpus construction as the dominant design leverSign in

  3. Distributed training patterns: parallelism as a systems boundary problemSign in

  4. Stability and schedules: making training predictable under scaleSign in

Part 18

Inference, Optimization, and Compression

  1. Inference, Optimization, and CompressionSign in

  2. Inference graphs and KV caching: throughput/latency trade-offs and streaming implicationsSign in

  3. Quantization, pruning, distillation: compressing capability into deployable footprintsSign in

  4. Hardware-aware deployment: CPUs, GPUs, accelerators, and where bottlenecks migrateSign in

Part 19

What Is a Foundation Model?

  1. What Is a Foundation Model?Sign in

  2. Self-Supervision and Emergence: why next-token and related objectives produce broad capabilities.Sign in

  3. Objective Families: masked, causal, and hybrid objectives as behavior-shaping choices.Sign in

  4. Scaling Laws Intuition: the data/compute/parameters triangle and what it predicts well (and poorly).Sign in

Part 20

LLMs and SLMs as Design Points

  1. LLMs and SLMs as Design PointsSign in

  2. LLM Design Posture: generality, capability, and the costs you accept to get them.Sign in

  3. SLM Design Posture: specialization, edge constraints, and operational privacy/cost advantages.Sign in

  4. Choosing a Point in the Trade Space: capability, latency, privacy, and unit economics.Sign in

Part 21

Fine-Tuning, Adaptation, and Instruction Following

  1. Fine-Tuning, Adaptation, and Instruction FollowingSign in

  2. Supervised Fine-Tuning: task shaping and the risks of brittle specialization.Sign in

  3. Adapters and LoRA-Style Methods: low-rank adaptation as an operationally tractable compromise.Sign in

  4. Instruction Tuning and Preference Data: steering behavior via curated interaction distributions.Sign in

Part 22

Alignment, Safety, and Guardrails

  1. Alignment, Safety, and GuardrailsSign in

  2. RLHF and Related Approaches: what is optimized, what is approximated, and where reward hacking appears.Sign in

  3. Red-Teaming and Safety Policies: turning anticipated misuse into testable evaluation artifacts.Sign in

  4. Guardrails and Constraints: balancing helpfulness, honesty, and harmlessness as a system design problem.Sign in

Part 23

Evaluation and Benchmarking of Foundation Models

  1. Evaluation and Benchmarking of Foundation ModelsSign in

  2. Benchmarks vs Holistic Evaluation: what benchmarks measure, and what they systematically miss.Sign in

  3. Robustness, Calibration, Bias, Fairness: evaluation as continuous monitoring of failure surfaces.Sign in

  4. Human Evaluation Loops: feedback pipelines, labeler variance, and governance of subjective judgments.Sign in

Part 24

Multi-Modal Architectures

  1. Multi-Modal ArchitecturesSign in

  2. Modality Encoders: vision, audio, code, and structured data as distinct input contracts.Sign in

  3. Fusion Strategies: cross-attention, late fusion, and joint embedding spaces.Sign in

  4. Contrastive and Joint Training: aligning modalities and the failure modes of misalignment.Sign in

Part 25

Retrieval-Augmented Generation (RAG)

  1. Retrieval-Augmented Generation (RAG)Sign in

  2. Vector Stores and Indexes: retrieval as an external memory boundary.Sign in

  3. Chunking and Context Construction: grounding as an engineering problem, not a slogan.Sign in

  4. RAG Failure Modes: hallucination, stale knowledge, retrieval miss, and robustness strategies.Sign in

Part 26

Tool-Using Models

  1. Tool-Using ModelsSign in

  2. Function Calling and Structured Outputs: schemas as the interface contract between models and tools.Sign in

  3. Designing Tool APIs for Models: affordances, constraints, and making actions safer than free-form text.Sign in

  4. Access Control and Safety Around Tools: permissions, auditing, and failure containment.Sign in

Part 27

Model Ecosystems and Composition

  1. Model Ecosystems and CompositionSign in

  2. Routing and Composition: orchestrating specialized models as a system-level mixture of experts.Sign in

  3. Specialized Helpers: code, math, extraction, planning, and when decomposition improves reliability.Sign in

  4. Latency, Reliability, and Cost: managing budgets and failure propagation in composed systems.Sign in

Part 28

Agents and Feedback Loops

  1. Agents and Feedback LoopsSign in

  2. Agent Loops: observe-think-act-learn as a control system, not just an app pattern.Sign in

  3. Planning and Reflection: when self-critique helps and when it creates new failure modes.Sign in

  4. Memory Architectures: short-term context versus long-term storage and retrieval boundaries.Sign in

Part 29

Multi-Agent Systems

  1. Multi-Agent SystemsSign in

  2. Coordination and Communication: protocols, roles, and division of labor among agents.Sign in

  3. Emergence and Simulation: what can arise from local rules and why it is hard to predict.Sign in

  4. Governance and Control: bounding agent societies with policies, incentives, and oversight.Sign in

Part 30

Orchestration Layers and Workflows

  1. Orchestration Layers and WorkflowsSign in

  2. Orchestration Frameworks and Policy Engines: separating execution from governance.Sign in

  3. Tool Routing, Retries, Backoff: reliability engineering for non-deterministic components.Sign in

  4. Integrating with Platforms and UIs: connecting agents to data, business systems, and human workflows.Sign in

Part 31

Operating AI Systems in Production

  1. Operating AI Systems in ProductionSign in

  2. Monitoring the Right Things: quality, safety, latency, and cost as co-equal SLOs.Sign in

  3. Incident Response for AI: rollbacks, kill switches, override mechanisms, and postmortems.Sign in

  4. Continuous Improvement: A/B testing, canaries, and closing the loop without contaminating evaluation.Sign in

Part 32

Organizational Design for AI Systems

  1. Organizational Design for AI SystemsSign in

  2. Roles and Interfaces: research, engineering, safety, product, and policy as coupled responsibilities.Sign in

  3. Data Lifecycle Governance: collection, labeling, privacy, and retention as design constraints.Sign in

  4. Documentation and Accountability: audits, model cards, and external reporting as operational necessities.Sign in

Part 33

World Models and Internal Simulators

  1. World Models and Internal SimulatorsSign in

  2. World Models in RL: latent dynamics and why prediction becomes simulation.Sign in

  3. Learning Environment Structure: representations of causality, dynamics, and uncertainty.Sign in

  4. Imagination and Counterfactuals: using simulators for planning and robustness.Sign in

Part 34

Planning and Control with Learned Models

  1. Planning and Control with Learned ModelsSign in

  2. Model Predictive Control: planning under constraints with learned dynamics.Sign in

  3. Integrating LMs with World Models: division of labor between language and dynamics.Sign in

  4. Closed-Loop Systems: perception -> model -> plan -> act as an end-to-end safety boundary.Sign in

Part 35

Continual Learning, Memory, and Identity

  1. Continual Learning, Memory, and IdentitySign in

  2. Lifelong Learning and Forgetting: catastrophic forgetting as a systems reliability problem.Sign in

  3. Memory Structures: internal vs external memory and how they change failure surfaces.Sign in

  4. Identity and Versioning Over Time: stability, updates, and continuity as product promises.Sign in

Part 36

AI in the Real World: Humans, Organizations, Society

  1. AI in the Real World: Humans, Organizations, SocietySign in

  2. Human-AI Interaction Patterns: interface design as behavior shaping and risk control.Sign in

  3. Societal Embedding: labor, creativity, decision-making, and organizational adoption dynamics.Sign in

  4. Governance and Long-Term Risk: regulation, accountability frameworks, and durable operational norms.Sign in

Part 37

Architectural Patterns for AI Models

  1. Architectural Patterns for AI ModelsSign in

  2. Encoder/Decoder Families: autoencoders and representation bottlenecks as reusable motifsSign in

  3. Diffusion and Energy-Based Perspectives: generative patterns and where they fit operationallySign in

  4. Hybrid Architectures: combining symbolic and neural components and the boundary management requiredSign in

  5. Architecture Selection by Step: matching pattern choice to ladder rung constraintsSign in

  6. Failure Surfaces by Architecture: typical breakdown modes and what to test firstSign in

Part 38

Data, Datasets, and Evaluation Patterns

  1. Data, Datasets, and Evaluation PatternsSign in

  2. Data Curation: filtering, deduplication, and provenance as model-shaping forcesSign in

  3. Privacy and Data Risk: sensitive data handling and the costs of leakageSign in

  4. Synthetic Data and Augmentation: when synthetic helps, and how it can misleadSign in

  5. Evaluation Suites and Dashboards: continuous measurement as part of the systemSign in

  6. Dataset Shift and Drift: maintaining meaning as environments and users changeSign in

Part 39

Infrastructure and Scaling Patterns

  1. Infrastructure and Scaling PatternsSign in

  2. Training Clusters: compute topology, throughput constraints, and distributed training patternsSign in

  3. Inference Infrastructure: online, batch, streaming, and the operational consequences of eachSign in

  4. Cost Optimization and Elasticity: scaling down as a first-class design goalSign in

  5. Hardware-Aware Design: choosing architectures and serving strategies that match acceleratorsSign in

  6. Reliability Under Load: degradation strategies when latency, cost, and quality collideSign in

Part 40

Safety, Security, and Robustness Patterns

  1. Safety, Security, and Robustness PatternsSign in

  2. Defense-in-Depth for AI: layered mitigations across model, data, and system boundariesSign in

  3. Adversarial Robustness and Abuse Resistance: anticipating malicious inputs and misuse incentivesSign in

  4. Policy, Logging, and Auditability: making failures diagnosable and decisions reviewableSign in

  5. Red-Teaming as an Engineering Practice: turning threats into test plans and regression suitesSign in

  6. Post-Hoc Analysis and Remediation: learning loops that improve safety without breaking trustSign in

Part 41

Design Patterns for AI Products

  1. Design Patterns for AI ProductsSign in

  2. Zero-Shot and Few-Shot UX: prompt-based interaction patterns and their brittlenessSign in

  3. Copilot, Agent, Assistant Archetypes: choosing product posture and responsibility boundariesSign in

  4. Human-in-the-Loop Workflows: escalation paths, approvals, and when automation should stopSign in

  5. Trust and Transparency Patterns: explaining uncertainty, citing sources, and managing expectationsSign in

  6. Product Safety in Practice: user reporting, abuse handling, and operational governanceSign in