"A class is complex when it can fit randomness on the sample it was handed."
Overview
Rademacher complexity is a data-dependent measure of how well a function class correlates with random signs.
Statistical learning theory is the chapter where finite samples meet future performance. It does not ask only whether a model performed well on observed data. It asks what assumptions let observed performance control unobserved risk.
This section is written in LaTeX Markdown. Inline mathematics uses $...$, and display equations use `
`. The notes emphasize theorem-level objects such as risk, hypothesis class, sample complexity, capacity, and confidence.
Prerequisites
Companion Notebooks
| Notebook | Description |
|---|---|
| theory.ipynb | Executable demonstrations for rademacher complexity |
| exercises.ipynb | Graded practice for rademacher complexity |
Learning Objectives
After completing this section, you will be able to:
- Define true risk, empirical risk, and generalization gap
- State PAC guarantees using and
- Derive finite-class sample complexity with union bounds
- Compute simple VC dimensions and growth functions
- Explain the bias-variance decomposition under squared loss
- Compare finite-class, VC, margin, norm, and Rademacher bounds
- Estimate simple Rademacher complexity by simulation
- Separate learnability theory from empirical benchmarking
- Identify when classical assumptions fail for modern AI systems
- Use learning-theory notation consistently with the repo guide
Study Flow
- Read the pages in order and pause after each page to restate the main definition or theorem.
- Run
theory.ipynbwhen you want to check the formulas numerically. - Use
exercises.ipynbafter the reading path, not before it. - Return to this overview page when you need the chapter-level navigation.