We want to talk about an inductor's error under a probabilistic supervised data model.
Let $(X, \mathcal{X} , \mu )$ and $f: X \to Y$ be a probabilistic supervised data model. Let $\mathcal{H} \subset (X \to Y)$ be a hypothesis class.
Let $\epsilon ,\delta \in (0, 1)$.
An inductor $A: (X \times Y)^n \to
\mathcal{H} $ is
$(\epsilon ,\delta )$-probably
approximately correct for $\mu $ and $f$
if
\[
\mu ^{n}\left[
\underset{\mu , f}{\mathword{error}}\left(
A((x_i, f(x_i))_{i = 1}^{n})
\right) \leq \epsilon
\right] \geq 1 - \delta .
\]
To show that an algorithm is probably approximately correct, we bound by $\delta $ the probability of the inductor producing a predictor which exceeds $\epsilon $ error.
A hypothesis class $\mathcal{H} $ is probably approximately correct learnable (or PAC learnable) if (a) for every underlying measure $\mu $, labeling function $f: X \to Y$, error $1-\epsilon $ and confidence $1-\delta $, (b) there exists $m_0 \in \N $ and a sequence of inductors $(A^m: (X \times Y)^{m} \to \mathcal{H} )_{m = m_0}^{\infty}$ so that (c) $A^m$ is $(\epsilon , \delta )$-probably approximately correct for all $m \geq m_0$ . If the inductors are subfamilies of a family of inductors $(A^n)_{n \in \N }$ then we say that this family of inductors PAC learns $\mathcal{H} $.
We interpret these conditions as follows. No matter the underlying distribution and labeling function, given an accuracy and confidence, we can specify the number of samples required to make the algorithm “succeed.” By which we mean that (“with high probability”) the predictor's error will be “small.”
If a family of inductors $\mathcal{A} = (A_n)_n$ PAC learns $\mathcal{H} $, then condition (b) above is equivalent to the existence function mapping $(\epsilon , \delta )$ to $\N $. Let $m_0^\star$ be the smallest integer such that condition (b) holds. We call the function $(\epsilon , \delta ) \mapsto m_0^\star(\epsilon ,\delta )$ the sample complexity of the family $\mathcal{A} $ on the hypothesis class $\mathcal{H} $.
For (what seem to be) historical reasons, some authors refer to the definition above as agnostic PAC learning since it drops a common condition (developed further in these sheets) relating the hypothesis class to the correct labeling function $f$.