This is an old revision of the document!
Hypothesis testing
Hypothesis testing involves deducing the quantity of a hypothesis $H$, which takes on one of the values $H_0, H_1, \dots$ from a measurement $R=r$.
Maximum //a posteriori// rule
We can do this by making the decision that minimizes the probability of error *conditional* on the measurement $R = r$.
- If $P(H_1|R = r) > P(H_0|R = r)$, that is, if it is more likely that $H = H_1$ than $H = H_0$ given that $R = r$, we decide '$H_1$'.
- Otherwise, if $P(H_1|R = r) < P(H_0|R = r)$, that is, if it is more likely that $H = H_1$ than $H = H_0$ given that $R = r$, we decide '$H_0$'.
The resulting conditional probability of error is:
$$ P(\mathrm{error}|R = r) = \min\{1 - P(H_0|R = r), 1 - P(H_1|R = r)\} $$
The conditional probabilities $P(H_1|R = r)$ and $P(H_0|R = r)$ are the a posteriori probabilities, as opposed to $P(H_1)$ and $P(H_0)$, the a priori probabilities.
The a posteriori probabilities can be calculated using Bayes' rule:
$$ P(H_0|R = r) = \frac{P(H_0) f_{R|H}(r|H_0)}{f_R(r)} $$
$$ P(H_1|R = r) = \frac{P(H_1) f_{R|H}(r|H_1)}{f_R(r)} $$
where $f_{R|H}$ is the conditional PDF of the random variable $R$ given a certain $H$, and $f_R$ is the PDF of $R$.
Since we are just comparing $P(H_0|R = r)$ and $P(H_1|R = r)$, we can cancel out the $f_R{r}$ on both sides, so it is equivalent to comparing $P(H_0) f_{R|H}(r|H_0)}$ and $P(H_1) f_{R|H}(r|H_1)}$:
- If $P(H_0) f_{R|H}(r|H_0)} > P(H_1) f_{R|H}(r|H_1)}$, then announce '$H_0$'.
- If $P(H_0) f_{R|H}(r|H_0)} < P(H_1) f_{R|H}(r|H_1)}$, then announce '$H_1$'.