Statistics basics

13 minute de lecture

Mis à jour :

1. Statistical modelling

TL; DR: We deal with identifiable models (i.e. different parameters lead to different distributions) with statistics (measure). In particular, we are interested in sufficient statistics, those that explain well the data. They verify have the advantage to statisfy the factorization theorem:

Another important characteristic is completeness: whether there is a unique unbiased estimator of $\theta$ based on the statistic (close to the Lehmann–Scheffé theorem)

Then we define exponential family which constitutes a nice framework as its canonical statistic is sufficient and complete, among others.

1.1. Generalities

In statistics, identifiability is a property which a model must satisfy in order for precise inference to be possible. It basically means that it is theoretically possible to learn the true values of this model’s underlying parameters after obtaining an infinite number of observations from it. Mathematically, it translates to:

Definition (Identifiability conditions): A model $(X, A , {P_{\theta} : \theta \in \Theta })$ is said identifiable if the mapping from $\Theta$ onto the probabilities space $(X, A)$, which to $\theta$ gives $P_{\theta}$ is injective.

This is equivalent to saying that different values of the parameters must generate different probability distributions of the observable variables.

Definition (Statistic) In a statistical model $(X, A , {P_{\theta} : \theta \in \Theta })$, a statistic is any measurable or $\sigma$-finite mapping $S$ from $(X, A)$ onto an arbitrary space. Let’s say a statistic is a function of r.V. $S(x_{1},…,x_{n})$.

1.2. Sufficiency

“…no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter.”

The formal definition of a sufficient statistic is a bit hard to grasp and can be difficult to use in practice:

where $L_{\theta}(X\mid S(X))$ is the conditional distribution of $X$ knowing $S(X)$. In consequence, one commonly refer to the Factorisation Criterion, which states that the statistic is sufficient it its likelihood functions is of the form:

Note: One could choose $S(X) = S$ of course, but the point is to find $S$ of minimal dimension.

1.3. Exponential family

A statistic $S$ is said to be complete if for any measurable real-valued function $\phi$, one has:

In essence, it ensures that the distributions corresponding to different values of the parameters are distinct. But once again, its hard to use in practice. Therefore, we will define family of models that is large enough to have expressive power and simple enough for theoretical manipulations. This family is the Exponential family and the condition for a model to belong to this family is that one can re-write its likelihood function as:

where $S(.) = (S_{1}(.), …, S_{r}(.))$ is the canonical statistic. The big advantage is that the canonical statistic is always sufficient and it is furthermore complete if $Q(\Theta)\neq\emptyset$.

For a given exponential model with $S_{i}$ are linearly independent in the affine sense, then:

This is quite powerful because it means that there is some sort of basis and that for two distributions to be equal, each of the component of the basis must be equal. The corollary is that under these conditions:

2. Unbiased estimation

In this section, we define regularity conditions ($C^{2}\cap L^{2}$ and support independent of $\theta$) on the likelihood function, and use them to build the Fisher Information Matrix (FIM), which measure of the amount of informations in the samples. This will be our basis to build bounds on the variance of an unbiased estimator (CRB) and optimal estimators for the parameters (Lemman-Scheffé). In particular, we will show that if $g(X)$ is any kind of estimator of a parameter $\theta$, then the conditional expectation of $g(X)$ given $T(X)$, where $T$ is a sufficient statistic, is typically a better estimator of $\theta$, and is never worse.

2.1. Regular model assumptions

  1. $\Theta$ is a an open set.
  2. The model is dominated by some measure
  3. The support $\Delta ={x\in H\mid L(x;\theta)>0}$ of the likelihood function is independend of $\theta$
  4. $L(x; \theta)$ is twice differentiable
  5. $L’$ and $L’’$ are integrable and we can permute the integral and the derivative.

2.2. Fisher Information

The score function $s_{\theta}(x)$ is simply the gradient of the log likelihood function defined by:

The score is zero-mean (i.e. $E[s_{\theta}(x)]=0$) and if it is square-integrable then we can define the FIM as the variance (covariance matrix in multidimensional case) of the score:

If we futhermore assume that the model is regular, then for any $\theta$:

The intuition behind the Fisher information is a way of measuring the amount of information that an observable random variable $X$ carries about an unknown parameter $\theta$ of a distribution that models $X$.

Practical note: In case of a $n$-sample, $(x_{1},…,x_{n})$ we have:

2.3. Optimality

The main idea is to give an answer $d$ regarding the data by using a loss function $\rho(d,\theta)$ between $d$ and the (true) value of the unknowns $\theta$ or $g(\theta)$. The loss function is usually quadratic;

with $A(.)$ is positive-definite. In this context, we define an estimator $g(\theta)$ as being a statistic $\delta(\textbf{x})$:

Finally, we can measure the error produced by the estimator using the Mean Squared Error:

2.4. Cramer Rao bound

The goal is to minimize the MSE but reformulated by the bias-variance dilemma

where $b_{\delta}(\theta)$ is the bias of $\delta(\textbf{x})$ (i.e. $b_{\delta}(\theta) = E_{\theta}[\delta(\textbf{x}) - g(\theta)]$).

In this subsection, we will give the following results:

  1. Rao-Blackwell estimator: An estimator with a sufficient statistic is always better than the same without (w.r.t. MSE).
  2. Lehman-Scheffé: If the statistic is additionally complete, it is the best of all unbiased estimators in terms of variance.
  3. Cramer-Rao lower Bound For a unbiased and regular estimator $\delta$ of $g \in C^{1}$, its variance is lower-bounded.

Theorem (Rao-Blackwell estimator) Let $\delta$ an estimator and $S$ a sufficient statistic. Let’s define an estimator with the knowledge of this statistic $\delta_{s} : x \mapsto E_{\theta}[\delta(\textbf{x}) \mid S(\textbf{x}) = S(x)]$. Then:

$\delta_{s}$ is Rao-Blackwell estimator (or the Rao-Blackwellization of $\delta$). As it turns out, it is unbiased if $\delta$ is unbiased.

Theorem: (Lehman-Scheffé optimality) If $\delta$ is unbiased and if $S$ is a sufficient and complete statistic, then the Rao-Blackwell estimator $\delta_{S}$ is optimal in the class of unbiased estimators, i.e. its variance is minimal for all $\theta \in \Theta$.

Definition (Regular estimator) $\delta$ is a regular estimator of the regular model $g(\theta)$ if:

  1. $E_{\theta}[\mid\delta \mid^{2}] < \infty\ \forall \theta \in \Theta$

  2. $\frac{\partial}{\partial\theta}\int_{X}\delta(x)l(x; \theta) dx= \int_{X}\delta(x)\frac{\partial}{\partial\theta}l(x; \theta) dx$

Theorem (Cramer-Rao lower Bound) Let $\delta$ an unbiased regular estimator of $g \in C^{1}(\mathbb{R}^{k})$ where $\theta \in \Theta \subset \mathbb{R}^{p}$. Let’s also assume that $I(\theta)$ be positive-definite. Then, for $n$-samples, and for all $\theta \in \Theta$, one has:

where $\frac{\partial g}{\partial \theta^{t}}(\theta)$ is the $p\times k$ matrix defined by: $(\frac{\partial g_{i}}{\partial \theta_{j}}(\theta))_{ij}$ and $\frac{\partial g^{t}}{\partial \theta}(\theta)$ its tranpose.

Note: In its simplest form, the bound states that the variance of any unbiased estimator is at least as high as the inverse of the Fisher information.

Definition (Efficiency): An unbiased estimator is said to be efficient if and only if its variance is the CRB.

Proposition: If $T$ is an efficient estimator of $g(\theta)$, then the affine transform $AT +b$ is an efficient estimator of $Ag(\theta)+b$ (for $A$ and $b$ with appropriate dimensions)

Proposition: An efficient estimator is optimal. The converse is (obviously) wrong.

The exponential model can be re-written in its natural form by making the change of variable $\lambda_{j} = Q_{h}(\theta)$:

where the new parameters are given by $(\lambda_{1},··· ,\lambda_{r}) \in \Lambda = Q(\Theta)\subset \mathbb{R}^{r}$

Theorem (Regularity) For an exponential and regular model with $\emptyset \neq \Lambda\subset \mathbb{R}^{r}$ that satisfy (A5) one obtain that $I(\lambda)$ and equals to:

Theorem (Identifiability): An exponential model with $\emptyset \neq \Lambda \subset \mathbb{R}^{r}$ is identifiable if and only if the FIM $I (\lambda)$ is invertible $\forall\lambda \in \Lambda$

Theorem (Necessary condition): Let’s consider a regular exponential model and $\delta$ an unbiased regular estimator $\delta$ of $g(\theta) \in C^{1}$. Moreover, let $I(\theta)$ be invertible $\forall \theta \in \Theta$. Thus, if $\delta$ is efficient, it is necessary an affine function of $S(x)=(S_{1}(x),···,S_{r}(x))$.

Theorem (Converse of CRB - equality) Given a regular model $g: \emptyset \neq \Theta \subset \mathbb{R}^{d} \mapsto \mathbb{R}^{p}$ of class $C^{1}$ such that

  • $\frac{\partial g}{\partial\theta^{t}(\theta)}$ is a square invertible matrix $\forall \theta \in \Theta$ so that $p = d$.

  • $I(\theta)$ exists and is invertible $\forall \theta \in \Theta$.

Then $\delta(x)$ is a regular and efficient estimator of $g(\theta)$ if and only if $L(x;\theta)$ can be written as:

where functions $Q$ and $C$ are s.t.

  • $Q$ and $C$ are differentiable $\forall \theta \in \Theta$
  • $\frac{\partial Q}{\partial\theta^{t}}(\theta)$ is invertible $\forall \theta \in \Theta$
  • $g(\theta) = -\big( \frac{\partial Q}{\partial\theta^{t}}(\theta) \big)^{-1}\frac{\partial ln(C)}{\partial\theta^{t}}(\theta$)

Corollary: In an exponential model of the natural form where $\emptyset \neq \Lambda \subset \mathbb{R}^{r}$ and where $I(\lambda)$ is invertible $\forall \lambda \in \Lambda$. Thus, each statistic $S_{j}(x)$ is an efficient estimator of $E_{\lambda}[S_{j}(X)]$ which is defined as:

3. Theory of point estimation

3.1. Basics

Definition (Consistancy) An estimator $\hat{\theta}{n}$ of $g(\theta)$ is strongly / weakly consistant if it $P{\theta_{0}}$ - almost surely / in proba converges torwards $g(\theta_{0})$, with $g: \Theta \mapsto \mathbb{R}^{p}$

Definition (Asymptotically unbiased) An estimator $\hat{\theta}_{n}$ of $g(\theta)$ is asymptotically unbiased if its limiting distribution is zero-mean, i.e.,

Remark: Different from “unbiased at the limit:” $E_{\theta_{0}}[\hat{\theta}{n}]\rightarrow g(\theta{0})$

Definition (Asymptotically normal) $\hat{\theta}_{n}$ os asymptotically normal if

where $\Sigma(\theta_{0})$ is the asymptotic CM of $\hat{\theta}_{n}$

Remark: This implies that $\hat{\theta}_{n}$ is asymptotically unbiased.

Definition (Asymptotically efficient): An estimator is asymptotically efficient if it is asymptotically normal and if:

3.2. Method of moment

Theorem

where $\textbf{U} = (U_{1}, …, U_{p})$ and $\textbf{m} = (m_{1}, …, m_{p})$

Theorem (Asymptotics of the MM estimator) If the function $\psi(\textbf{U}) = \theta_{n}$ is differentiable then:

where $\textbf{A}(\theta) = \frac{\partial\psi}{\partial\theta^{t}}(m) \Sigma(\theta) \frac{\partial\psi^{t}}{\partial\theta}(m)$

MME strongly consistant, asymptotically normal BUT generally NOT asymptotically efficient!

3.3. Method of Maximum Likelihood

Definition (Maximum Likelihood Estimator (MLE))

Consider a regular model that satisfy (A5) and (A6)

(A6) $\forall x \in \Delta$, for $\theta$ close to $\theta_{0}$, $\text{log}(f(x;\theta))$ is three times differentiable w.r.t. $\theta$ and

with $E_{\theta_{0}}[M(x)] < + \infty$

Proposition: Assume that the model is identifiable then $\forall \theta \neq \theta_{0}$, one has:

Definition (MLE) The MLE is defined by

The MLE has to verified the following likelihood equations!

Proposition: Let $g:\Theta\mapsto\mathbb{R}^{p}$. If $\hat{\theta}{n}$ is the MLE of $\theta$ then $g(\hat{\theta}{n})$ is also a MLE of $g(\theta)$

Note: The MLE is not necessary unique..

Theorem: Assume that the model is identifiable, that (A1) and (A2) hold and that $\theta_{0} \in \Theta \neq \emptyset$ compact and:

  • $x_{1} \mapsto L(x_{1}, \theta)$ is bounded $\forall \theta \in \Theta$
  • $\theta \mapsto L(x_{1}, \theta)$ is continuous $\forall x_{1} \in \Delta$

Theorem (Classical asymptotics)

Assume that the model is identifiable and that $\Theta$ is an open set of $\mathbb{R}^{d}$ and (A1)-(A6) hold. Then, it exists $\hat{\theta}{n}^{ML}$ (from a given $n{0}$) solution to the likelihood equations such that:

If furthermore we have that $g$ is differentiable then it exists $\hat{\theta}{n}^{ML}$ (from a given $n{0}$) solution to the likelihood equations such that:

Conclusions: The MLE is strongly consistant, asymptotically normal and asymptotically efficient.

Come back on exponential models Let an exponential model with $\lambda \in \Lambda \neq \emptyset$ and $I(\lambda)$ invertible $\forall \lambda in \Lambda$ (identifiable model). Thus, the MLE exists (from a given n_{0}), is unique, strongly consistant and asymptotically efficient (which includes asymptotically normal).

3.4. Bayesian estimation

Principles: Philosophy is different from previous MM/ML estimation approaches (frequentist methods). The purpose is the same: estimating an unknown parameter $\theta \in \mathbb{R}$ or $\mathbb{R}^{p}$ thanks to the sample $(x_{1},…,x_{n})$ likelihood (parameterized by $\theta$) and an a priori distribution $p(\theta)$. So, $\theta$ is assumed to random.

Ideas: To that end, one has to minimize a cost function $c(\theta,\hat{\theta})$ that represents the error between $\theta$ and its estimator $\hat{\theta}$.

Reminders: A posteriori distribution / posterior distribution

MMSE estimator (mean of the posterior PDF) is the estimator that minimizes the MSE as the cost function:

There are two cases:

  1. $\theta \in \mathbb{R}$

so

  1. $\theta \in \mathbb{R}^{p}$. The MMSE estimator $\hat{\theta}_{MMSE}(x) = E[\theta \mid x]$ minimizes the quadratic cost:

for any symmetric definite positive matrix $Q$ (and in particular for $Q = I_{p}$ the identity matrix).

MAP estimator: Once again there are two cases:

  1. $\theta \in \mathbb{R}$

The MAP estimator $\hat{\theta}_{MAP}(\textbf{x})$ minimizes the average of a uniform cost function:

and is defined by:

If $\Lambda$ is arbitrarily small, $\hat{\theta}_{MAP}(\textbf{x})$ is the value of $\pi(x)$ which maximizes the posterior $p(\theta\mid x)$ hence its name MAP estimator. It is computed by setting to zero the derivative of $p(\theta \mid x)$ or its log with respect to $\theta$.

  1. $\theta \in \mathbb{R}^{d}$

Determine the values of $\theta_{j}$ which make the partial derivatives of $p(\theta \mid \textbf{x})$ (or its logarithm) with respect to $\theta_{i}$ equal to zero.


Laisser un commentaire