kb:estimation_methods

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
kb:estimation_methods [2022-02-12 18:38] – [Feature matching] jaeyoungkb:estimation_methods [2024-04-30 04:03] (current) – external edit 127.0.0.1
Line 36: Line 36:
 For a given probability distribution $\mathbb{P}$ with parameter $\theta$, we can extract feature(s) $h^\theta = g(\mathbb{P}^\theta)$. We can also calculate the features for the empirical distribution $\hat{h} = g(\hat{\mathbb{P}})$. Then solve for $\theta$ by setting $h^\theta = \hat{h}$. For a given probability distribution $\mathbb{P}$ with parameter $\theta$, we can extract feature(s) $h^\theta = g(\mathbb{P}^\theta)$. We can also calculate the features for the empirical distribution $\hat{h} = g(\hat{\mathbb{P}})$. Then solve for $\theta$ by setting $h^\theta = \hat{h}$.
  
-For examplefind the median $\hat{h}$ for sampled data, and find $\theta$ such that the median of the estimated distribution $h^\theta$ is equal to $\hat{h}$.+==== Method of moments ==== 
 + 
 +Moments of distributions are commonly used as features for feature matching. The $k$-th moment of a random variable $X$ is $\mathbb{E}[X^k]$. 
 + 
 +To estimate the moment from empirical data $X_1... X_n$, replace expectation with the average: 
 + 
 +$$ \hat{\mathbb{E}}[X^k] = \frac{1}{n} \sum_{i=1}^n X_i^k $
 +===== Maximum likelihood estimator ===== 
 + 
 +Assume a probability mass or distribution function with parameter(s) $\theta$. Given a set of data points $ X = (X_1, ..., X_n) $, the likelihood function is the product of the PMFs of all of the points for a discrete distribution, or the product of the PDFs of all of the points for a continuous distribution. 
 + 
 +Discrete (PMF): 
 + 
 +$L^\theta(x_1, ..., x_n) = \prod_{i=1}^{n} \mathbb{P}^\theta (X_i = x_i) $$ 
 + 
 +Continuous (PDF): 
 + 
 +$$ L^\theta(x_1, ..., x_n) = \prod_{i=1}^{n} f_{X_i}^\theta (x_i) $$ 
 + 
 +==== Log-likelihood ==== 
 + 
 +It is usually easier to maximize the log of the likelihood function, known as the log-likelihood function. This is of course equivalent to maximizing the likelihood function. 
 + 
 +Discrete (PMF): 
 + 
 +$$ \max_\theta \sum_{i=1}^n \log \mathbb{P}^\theta (X_i = x_i) $$ 
 + 
 +Continuous (PDF): 
 + 
 +$$ \max_\theta \sum_{i=1}^n \log f_{X_i}^\theta (x_i) $$
  • kb/estimation_methods.1644691098.txt.gz
  • Last modified: 2024-04-30 04:03
  • (external edit)